Prilepsky, Jaroslaw E.
Computational Complexity Evaluation of Neural Network Applications in Signal Processing
Freire, Pedro, Srivallapanondh, Sasipim, Napoli, Antonio, Prilepsky, Jaroslaw E., Turitsyn, Sergei K.
In this paper, we provide a systematic approach for assessing and comparing the computational complexity of neural network layers in digital signal processing. We provide and link four software-to-hardware complexity measures, defining how the different complexity metrics relate to the layers' hyper-parameters. This paper explains how to compute these four metrics for feed-forward and recurrent layers, and defines in which case we ought to use a particular metric depending on whether we characterize a more soft- or hardware-oriented application. One of the four metrics, called `the number of additions and bit shifts (NABS)', is newly introduced for heterogeneous quantization. NABS characterizes the impact of not only the bitwidth used in the operation but also the type of quantization used in the arithmetical operations. We intend this work to serve as a baseline for the different levels (purposes) of complexity estimation related to the neural networks' application in real-time digital signal processing, aiming at unifying the computational complexity estimation.
Hardware Realization of Nonlinear Activation Functions for NN-based Optical Equalizers
Srivallapanondh, Sasipim, Freire, Pedro J., Napoli, Antonio, Turitsyn, Sergei K., Prilepsky, Jaroslaw E.
Thus, different digital signal processing (DSP) techniques have been proposed to alleviate nonlinear effects [1]. More recently, the research focus has shifted to the deployment of neural networks (NNs) for optical channel post-equalization, because NNs have shown the capability to estimate the inverse optical channel transfer function and successfully mitigate the nonlinearity. However, major challenges still prevent NN-based equalizers from being implemented in real hardware, in particular, the high computational complexity [2-4]. The complexity of nonlinear activation functions is one of the key components when designing the NNs in hardware. In the hardware realization, the weights and inputs of the NN can be converted directly from the float-to fixed-point representation, while the nonlinear activation function's realization is more challenging.
Implementing Neural Network-Based Equalizers in a Coherent Optical Transmission System Using Field-Programmable Gate Arrays
Freire, Pedro J., Srivallapanondh, Sasipim, Anderson, Michael, Spinnler, Bernhard, Bex, Thomas, Eriksson, Tobias A., Napoli, Antonio, Schairer, Wolfgang, Costa, Nelson, Blott, Michaela, Turitsyn, Sergei K., Prilepsky, Jaroslaw E.
In this work, we demonstrate the offline FPGA realization of both recurrent and feedforward neural network (NN)-based equalizers for nonlinearity compensation in coherent optical transmission systems. First, we present a realization pipeline showing the conversion of the models from Python libraries to the FPGA chip synthesis and implementation. Then, we review the main alternatives for the hardware implementation of nonlinear activation functions. The main results are divided into three parts: a performance comparison, an analysis of how activation functions are implemented, and a report on the complexity of the hardware. The performance in Q-factor is presented for the cases of bidirectional long-short-term memory coupled with convolutional NN (biLSTM + CNN) equalizer, CNN equalizer, and standard 1-StpS digital back-propagation (DBP) for the simulation and experiment propagation of a single channel dual-polarization (SC-DP) 16QAM at 34 GBd along 17x70km of LEAF. The biLSTM+CNN equalizer provides a similar result to DBP and a 1.7 dB Q-factor gain compared with the chromatic dispersion compensation baseline in the experimental dataset. After that, we assess the Q-factor and the impact of hardware utilization when approximating the activation functions of NN using Taylor series, piecewise linear, and look-up table (LUT) approximations. We also show how to mitigate the approximation errors with extra training and provide some insights into possible gradient problems in the LUT approximation. Finally, to evaluate the complexity of hardware implementation to achieve 200G and 400G throughput, fixed-point NN-based equalizers with approximated activation functions are developed and implemented in an FPGA.
Knowledge Distillation Applied to Optical Channel Equalization: Solving the Parallelization Problem of Recurrent Connection
Srivallapanondh, Sasipim, Freire, Pedro J., Spinnler, Bernhard, Costa, Nelson, Napoli, Antonio, Turitsyn, Sergei K., Prilepsky, Jaroslaw E.
Moreover, with the ever-increasing transmission bandwidth, nonlinearity becomes even more important [1]. Various digital signal processing (DSP) techniques have been proposed to minimize nonlinear effects [2]. Due to the universal approximation capability of neural networks (NNs), the NNs have recently been intensively studied for the optical channel post-equalization, because they can approximate the inverse optical channel transfer function with good accuracy and revert the nonlinear distortions. In particular, recurrent NNs (RNN) based equalizers have shown the best capability in equalizing nonlinear impairments as compared to the feed-forward NN types [3-5]. However, since the RNN structure has a feedback loop, it is not easily parallelizable.
Reducing Computational Complexity of Neural Networks in Optical Channel Equalization: From Concepts to Implementation
Freire, Pedro J., Napoli, Antonio, Ron, Diego Arguello, Spinnler, Bernhard, Anderson, Michael, Schairer, Wolfgang, Bex, Thomas, Costa, Nelson, Turitsyn, Sergei K., Prilepsky, Jaroslaw E.
In this paper, a new methodology is proposed that allows for the low-complexity development of neural network (NN) based equalizers for the mitigation of impairments in high-speed coherent optical transmission systems. In this work, we provide a comprehensive description and comparison of various deep model compression approaches that have been applied to feed-forward and recurrent NN designs. Additionally, we evaluate the influence these strategies have on the performance of each NN equalizer. Quantization, weight clustering, pruning, and other cutting-edge strategies for model compression are taken into consideration. In this work, we propose and evaluate a Bayesian optimization-assisted compression, in which the hyperparameters of the compression are chosen to simultaneously reduce complexity and improve performance. In conclusion, the trade-off between the complexity of each compression approach and its performance is evaluated by utilizing both simulated and experimental data in order to complete the analysis. By utilizing optimal compression approaches, we show that it is possible to design an NN-based equalizer that is simpler to implement and has better performance than the conventional digital back-propagation (DBP) equalizer with only one step per span. This is accomplished by reducing the number of multipliers used in the NN equalizer after applying the weighted clustering and pruning algorithms. Furthermore, we demonstrate that an equalizer based on NN can also achieve superior performance while still maintaining the same degree of complexity as the full electronic chromatic dispersion compensation block. We conclude our analysis by highlighting open questions and existing challenges, as well as possible future research directions.