Goto

Collaborating Authors

 Dörner, Sebastian


Deep Reinforcement Learning for mmWave Initial Beam Alignment

arXiv.org Artificial Intelligence

We investigate the applicability of deep reinforcement learning algorithms to the adaptive initial access beam alignment problem for mmWave communications using the state-of-the-art proximal policy optimization algorithm as an example. In comparison to recent unsupervised learning based approaches developed to tackle this problem, deep reinforcement learning has the potential to address a new and wider range of applications, since, in principle, no (differentiable) model of the channel and/or the whole system is required for training, and only agent-environment interactions are necessary to learn an algorithm (be it online or using a recorded dataset). We show that, although the chosen off-the-shelf deep reinforcement learning agent fails to perform well when trained on realistic problem sizes, introducing action space shaping in the form of beamforming modules vastly improves the performance, without sacrificing much generalizability. Using this add-on, the agent is able to deliver competitive performance to various state-of-the-art methods on simulated environments, even under realistic problem sizes. This demonstrates that through well-directed modification, deep reinforcement learning may have a chance to compete with other approaches in this area, opening up many straightforward extensions to other/similar scenarios.


Enabling FDD Massive MIMO through Deep Learning-based Channel Prediction

arXiv.org Machine Learning

A major obstacle for widespread deployment of frequency division duplex (FDD)-based Massive multiple-input multipleoutput (MIMO)communications is the large signaling overhead for reporting full downlink (DL) channel state information (CSI) back to the basestation (BS), in order to enable closed-loop precoding. We completely remove this overhead by a deep-learning based channel extrapolation (or "prediction") approach and demonstrate that a neural network (NN) at the BS can infer the DL CSI centered around a frequency f UL; nomore pilot/reporting overhead is needed than with a genuine time division duplex (TDD)-based system. The rationale is that scatterers and the large-scale propagation environment are sufficiently similar to allow a NN to learn about the physical connections and constraints between two neighboring frequency bands, and thus provide a well-operating system even when classic extrapolation methods, like the Wiener filter (used as a baseline for comparison throughout) fails. We study its performance for various state-of-the-art Massive MIMO channel models, and, even more so, evaluate the scheme using actual Massive MIMO channel measurements, rendering it to be practically feasible at negligible loss in spectral efficiency when compared to a genuine TDD-based system. I. INTRODUCTION With a significant increase in area throughput, Massive multiple-input multiple-output (MIMO) antenna communication has become an enabling technology for the upcoming fifth generation (5G) wireless mobile communication systems [1], [2], [3], [4]. However, Massive MIMO systems described in current research literature commonly exploit channel reciprocity and hence rely on time division duplex (TDD)-based approaches [1], i.e., uplink (UL) and downlink (DL) channels share the same frequency band in orthogonal time intervals. Achieving such reciprocity in practice requires accurate hardware with costly calibration circuitry. To mitigate this issue, various alternatives to a full Massive MIMO system have been proposed such as the grid of beams [5] and codebook Massive MIMO [6].


Deep Learning-Based Communication Over the Air

arXiv.org Machine Learning

End-to-end learning of communications systems is a fascinating novel concept that has so far only been validated by simulations for block-based transmissions. It allows learning of transmitter and receiver implementations as deep neural networks (NNs) that are optimized for an arbitrary differentiable end-to-end performance metric, e.g., block error rate (BLER). In this paper, we demonstrate that over-the-air transmissions are possible: We build, train, and run a complete communications system solely composed of NNs using unsynchronized off-the-shelf software-defined radios (SDRs) and open-source deep learning (DL) software libraries. We extend the existing ideas towards continuous data transmission which eases their current restriction to short block lengths but also entails the issue of receiver synchronization. We overcome this problem by introducing a frame synchronization module based on another NN. A comparison of the BLER performance of the "learned" system with that of a practical baseline shows competitive performance close to 1 dB, even without extensive hyperparameter tuning. We identify several practical challenges of training such a system over actual channels, in particular the missing channel gradient, and propose a two-step learning procedure based on the idea of transfer learning that circumvents this issue.