Deep Compression of Sum-Product Networks on Tensor Networks
Ko, Ching-Yun, Chen, Cong, Zhang, Yuke, Batselier, Kim, Wong, Ngai
Abstract--Sum-product networks (SPNs) represent an emerging class of neural networks with clear probabilistic semantics and superior inference speed over graphical models. This work reveals a strikingly intimate connection between SPNs and tensor networks, thus leading to a highly efficient representation that we call tensor SPNs (tSPNs). For the first time, through mapping an SPN onto a tSPN and employing novel optimization techniques, we demonstrate remarkable parameter compression with negligible loss in accuracy. INCE the inception of sum-product networks (SPNs) [1], a multitude of works have emerged with respect to their structure and weight learning, e.g., [2], [3], [4], as well as their application in image completion, speech modeling, semantic mapping and robotics, e.g., [5], just to name a few. An SPN exhibits a clear semantics of mixtures (sum nodes) and features (product nodes). Compared to other probabilistic graphical models like Bayesian and Markov networks with #P or NPhard computation, an SPN enjoys a tractable exact inference cost, and its learning is relatively simple and fast. On the other hand, there has been an exploding number of works on tensors (a multilinear operator rooted in physics) [6] including their connection and utilization in various engineering fields such as signal processing [7], and lately also in neural networks and machine learning [8], [9], [10].
Nov-9-2018