Controlling the Complexity of HMM Systems by Regularization
Neukirchen, Christoph, Rigoll, Gerhard
–Neural Information Processing Systems
This paper introduces a method for regularization ofHMM systems that avoids parameter overfitting caused by insufficient training data. Regularization isdone by augmenting the EM training method by a penalty term that favors simple and smooth HMM systems. The penalty term is constructed as a mixture model of negative exponential distributions that is assumed to generate the state dependent emission probabilities of the HMMs. This new method is the successful transfer of a well known regularization approach in neural networks to the HMM domain and can be interpreted as a generalization of traditional state-tying for HMM systems. Theeffect of regularization is demonstrated for continuous speech recognition tasks by improving overfitted triphone models and by speaker adaptation with limited training data. 1 Introduction One general problem when constructing statistical pattern recognition systems is to ensure the capability to generalize well, i.e. the system must be able to classify data that is not contained in the training data set.
Neural Information Processing Systems
Dec-31-1999