Kalman Normalization: Normalizing Internal Representations Across Network Layers

Neural Information Processing Systems 

As an indispensable component, Batch Normalization (BN) has successfully improved the training of deep neural networks (DNNs) with mini-batches, by normalizing the distribution of the internal representation for each hidden layer.