Skalli, Anas
Model-free front-to-end training of a large high performance laser neural network
Skalli, Anas, Sunada, Satoshi, Goldmann, Mirko, Gebski, Marcin, Reitzenstein, Stephan, Lott, James A., Czyszanowski, Tomasz, Brunner, Daniel
Artificial neural networks (ANNs), have become ubiquitous and revolutionized many applications ranging from computer vision to medical diagnoses. However, they offer a fundamentally connectionist and distributed approach to computing, in stark contrast to classical computers that use the von Neumann architecture. This distinction has sparked renewed interest in developing unconventional hardware to support more efficient implementations of ANNs, rather than merely emulating them on traditional systems. Photonics stands out as a particularly promising platform, providing scalability, high speed, energy efficiency, and the ability for parallel information processing. However, fully realized autonomous optical neural networks (ONNs) with in-situ learning capabilities are still rare. In this work, we demonstrate a fully autonomous and parallel ONN using a multimode vertical cavity surface emitting laser (VCSEL) using off-the-shelf components. Our ONN is highly efficient and is scalable both in network size and inference bandwidth towards the GHz range. High performance hardware-compatible optimization algorithms are necessary in order to minimize reliance on external von Neumann computers to fully exploit the potential of ONNs. As such we present and extensively study several algorithms which are broadly compatible with a wide range of systems. We then apply these algorithms to optimize our ONN, and benchmark them using the MNIST dataset. We show that our ONN can achieve high accuracy and convergence efficiency, even under limited hardware resources. Crucially, we compare these different algorithms in terms of scaling and optimization efficiency in term of convergence time which is crucial when working with limited external resources. Our work provides some guidance for the design of future ONNs as well as a simple and flexible way to train them.
Limits of nonlinear and dispersive fiber propagation for photonic extreme learning
Ermolaev, Andrei V., Hary, Mathilde, Leybov, Lev, Ryczkowski, Piotr, Skalli, Anas, Brunner, Daniel, Genty, Goëry, Dudley, John M.
We report a generalized nonlinear Schr\"odinger equation simulation model of an extreme learning machine (ELM) based on optical fiber propagation. Using handwritten digit classification as a benchmark, we study how accuracy depends on propagation dynamics, as well as parameters governing spectral encoding, readout, and noise. Test accuracies of over 91% and 93% are found for propagation in the anomalous and normal dispersion regimes respectively. Our simulation results also suggest that quantum noise on the input pulses introduces an intrinsic penalty to ELM performance.
Training of Physical Neural Networks
Momeni, Ali, Rahmani, Babak, Scellier, Benjamin, Wright, Logan G., McMahon, Peter L., Wanjura, Clara C., Li, Yuhang, Skalli, Anas, Berloff, Natalia G., Onodera, Tatsuhiro, Oguz, Ilker, Morichetti, Francesco, del Hougne, Philipp, Gallo, Manuel Le, Sebastian, Abu, Mirhoseini, Azalia, Zhang, Cheng, Marković, Danijela, Brunner, Daniel, Moser, Christophe, Gigan, Sylvain, Marquardt, Florian, Ozcan, Aydogan, Grollier, Julie, Liu, Andrea J., Psaltis, Demetri, Alù, Andrea, Fleury, Romain
Physical neural networks (PNNs) are a class of neural-like networks that leverage the properties of physical systems to perform computation. While PNNs are so far a niche research area with small-scale laboratory demonstrations, they are arguably one of the most underappreciated important opportunities in modern AI. Could we train AI models 1000x larger than current ones? Could we do this and also have them perform inference locally and privately on edge devices, such as smartphones or sensors? Research over the past few years has shown that the answer to all these questions is likely "yes, with enough research": PNNs could one day radically change what is possible and practical for AI systems. To do this will however require rethinking both how AI models work, and how they are trained - primarily by considering the problems through the constraints of the underlying hardware physics. To train PNNs at large scale, many methods including backpropagation-based and backpropagation-free approaches are now being explored. These methods have various trade-offs, and so far no method has been shown to scale to the same scale and performance as the backpropagation algorithm widely used in deep learning today. However, this is rapidly changing, and a diverse ecosystem of training techniques provides clues for how PNNs may one day be utilized to create both more efficient realizations of current-scale AI models, and to enable unprecedented-scale models.