Goto

Collaborating Authors

 hinton


Self-Routing Capsule Networks

Taeyoung Hahn, Myeongjang Pyeon, Gunhee Kim

Neural Information Processing Systems

In this work, we propose a novel and surprisingly simple routing strategy called self-routing, where each capsule is routed independently by its subordinate routing network. Therefore, the agreement between capsules is not required anymore, but both poses and activations of upper-level capsules are obtained in a way similar to Mixture-of-Experts. Our experiments on CIFAR10, SVHN, and SmallNORB showthat the self-routing performs more robustly against white-box adversarial attacks and affine transformations, requiring less computation.




Deep Neural Nets with Interpolating Function as Output Activation

Bao Wang, Xiyang Luo, Zhen Li, Wei Zhu, Zuoqiang Shi, Stanley Osher

Neural Information Processing Systems

And we propose end-to-end training and testing algorithms for this new architecture. Compared to classical neural nets with softmax function as output activation, the surrogate with interpolating function as output activation combines advantages of both deep and manifold learning.



Legendre Decomposition for Tensors

Mahito Sugiyama, Hiroyuki Nakahara, Koji Tsuda

Neural Information Processing Systems

CP decomposition compresses an input tensor into a sum of rank-one components, and Tucker decomposition approximates an input tensor by a core tensor multiplied by matrices. To date, matrix and tensor decomposition has been extensively analyzed, and there are a number of variations of such decomposition (Kolda and Bader, 2009), where the common goal is to approximate a given tensor by a smaller number of components, or parameters,inanefficientmanner. However, despite the recent advances of decomposition techniques, a learning theory that can systematically define decomposition for any order tensors including vectors and matrices is still under development. Moreover, it is well known that CP and Tucker tensor decomposition include non-convex optimization and that the global convergence is not guaranteed.





Stacked Capsule Autoencoders

Adam Kosiorek, Sara Sabour, Yee Whye Teh, Geoffrey E. Hinton

Neural Information Processing Systems

Objects are composed of a set of geometrically organized parts. We introducean unsupervised capsule autoencoder ( SCAE), which explicitly uses geometric relationships between parts toreason about objects. Since these relationships do not depend on the viewpoint, our model is robust to viewpoint changes.