Near-Optimal Smoothing of Structured Conditional Probability Matrices Mesrob I. Ohannessian University of California, San Diego Toyota Technological Institute at Chicago San Diego, CA, USA
–Neural Information Processing Systems
Utilizing the structure of a probabilistic model can significantly increase its learning speed. Motivated by several recent applications, in particular bigram models in language processing, we consider learning low-rank conditional probability matrices under expected KL-risk. This choice makes smoothing, that is the careful handling of low-probability elements, paramount. We derive an iterative algorithm that extends classical non-negative matrix factorization to naturally incorporate additive smoothing and prove that it converges to the stationary points of a penalized empirical risk. We then derive sample-complexity bounds for the global minimzer of the penalized risk and show that it is within a small factor of the optimal sample complexity.
Neural Information Processing Systems
Mar-12-2024, 14:15:01 GMT
- Country:
- North America > United States > California > San Diego County > San Diego (0.77)
- Industry:
- Automobiles & Trucks > Manufacturer (0.40)