Goto

Collaborating Authors

 Brooks, Daniel


Riemannian batch normalization for SPD neural networks

arXiv.org Machine Learning

Covariance matrices have attracted attention for machine learning applications due to their capacity to capture interesting structure in the data. The main challenge is that one needs to take into account the particular geometry of the Riemannian manifold of symmetric positive definite (SPD) matrices they belong to. In the context of deep networks, several architectures for these matrices have recently been proposed. In our article, we introduce a Riemannian batch normalization (batchnorm) algorithm, which generalizes the one used in Euclidean nets. This novel layer makes use of geometric operations on the manifold, notably the Riemannian barycenter, parallel transport and non-linear structured matrix transformations. We derive a new manifold-constrained gradient descent algorithm working in the space of SPD matrices, allowing to learn the batchnorm layer. We validate our proposed approach with experiments in three different contexts on diverse data types: a drone recognition dataset from radar observations, and on emotion and action recognition datasets from video and motion capture data. Experiments show that the Riemannian batchnorm systematically gives better classification performance compared with leading methods and a remarkable robustness to lack of data.


Towards State Summarization for Autonomous Robots

AAAI Conferences

Mobile robots are an increasingly important part of search and rescue efforts as well as military combat. 
In order for users to accept these robots and use them effectively, the user must be able to communicate clearly with the robots and obtain explanations of the robots' behavior that will allow the user to understand its actions. 
This paper describes part of a system of software that will be able to produce explanations of the robots' behavior and situation in an interaction with a human operator.