Multifidelity Kolmogorov-Arnold Networks

Howard, Amanda A., Jacob, Bruno, Stinis, Panos

arXiv.org Artificial Intelligence 

In recent years, scientific machine learning (SciML) has emerged as a paradigm for modeling physical systems [1, 2, 3]. Typically using the theory of multilayer perceptrons (MLPs), SciML has shown great success in modeling a wide range of applications, however, data-informed training struggles when high-quality data is not available. Kolmogorov-Arnold networks (KANs) have recently been developed as an alternative to MLPs [4, 5]. KANs use the Kolmogorov-Arnold Theorem as inspiration and can offer advantages over MLPs in some cases, such as for discovering interpretable models. However, KANs have been shown to struggle to reach the accuracy of MLPs, particularly without modifications [6, 7, 8, 9]. In the short time since the publication of [4], many variations of KANs have been developed, including physics-informed KANs (PIKANs)[9], KAN-informed neural networks (KINNs)[10], temporal KANs [11], wavelet KANs [12], graph KANs [13, 14, 15], Chebyshev KANs (cKANs) [16], convolutional KANs [17], ReLU-KANs [18], Higher-order-ReLU-KANs (HRKANs) [19], fractional KANs [20], finite basis KANs [21], deep operator KANs [22], and others.