Goto

Collaborating Authors

 matrix chernoff bound


A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices

Neural Information Processing Systems

We prove a Chernoff-type bound for sums of matrix-valued random variables sampled via a regular (aperiodic and irreducible) finite Markov chain. Specially, consider a random walk on a regular Markov chain and a Hermitian matrix-valued function on its state space. Our result gives exponentially decreasing bounds on the tail distributions of the extreme eigenvalues of the sample mean matrix. Our proof is based on the matrix expander (regular undirected graph) Chernoff bound [Garg et al.


Review for NeurIPS paper: A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices

Neural Information Processing Systems

Summary and Contributions: This paper establishes a concentration inequality in operator norm for the sample co-occurrence matrix C of a regular finite-state Markov chain. This is the matrix whose (i,j) entry counts the fraction of times states i and j co-occur within a time window of fixed size T, with a potential weighting by the difference of their occurrence times. The probability that C-E[C] exceeds eps is shown to be exponentially small in eps 2 * L, where L is the sample size (i.e. the total length of the chain). This concentration inequality is established as a corollary of a general result for the concentration of sample averages for a bounded symmetric-matrix-valued function applied to samples from an ergodic Markov chain. This general result is of independent interest.


Review for NeurIPS paper: A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices

Neural Information Processing Systems

This work tackles the problem of estimating the co-occurrence matrix of a Markov chain. The referees were unanimous in the assessment that this is a solid contribution, worthy of being accepted. The only reservations were some missing references in related work, which the authors agreed to discuss in the revision.


A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices

Neural Information Processing Systems

We prove a Chernoff-type bound for sums of matrix-valued random variables sampled via a regular (aperiodic and irreducible) finite Markov chain. Specially, consider a random walk on a regular Markov chain and a Hermitian matrix-valued function on its state space. Our result gives exponentially decreasing bounds on the tail distributions of the extreme eigenvalues of the sample mean matrix. Our proof is based on the matrix expander (regular undirected graph) Chernoff bound [Garg et al. Our matrix Chernoff bound for Markov chains can be applied to analyze the behavior of co-occurrence statistics for sequential data, which have been common and important data signals in machine learning.


A Matrix Chernoff Bound for Markov Chains and Its Application to Co-occurrence Matrices

Qiu, Jiezhong, Wang, Chi, Liao, Ben, Peng, Richard, Tang, Jie

arXiv.org Machine Learning

We prove a Chernoff-type bound for sums of matrix-valued random variables sampled via a regular (aperiodic and irreducible) finite Markov chain. Specially, consider a random walk on a regular Markov chain and a Hermitian matrix-valued function on its state space. Our result gives exponentially decreasing bounds on the tail distributions of the extreme eigenvalues of the sample mean matrix. Our proof is based on the matrix expander (regular undirected graph) Chernoff bound [Garg et al. STOC '18] and scalar Chernoff-Hoeffding bounds for Markov chains [Chung et al. STACS '12]. Our matrix Chernoff bound for Markov chains can be applied to analyze the behavior of co-occurrence statistics for sequential data, which have been common and important data signals in machine learning. We show that given a regular Markov chain with $n$ states and mixing time $\tau$, we need a trajectory of length $O(\tau (\log{(n)}+\log{(\tau)})/\epsilon^2)$ to achieve an estimator of the co-occurrence matrix with error bound $\epsilon$. We conduct several experiments and the experimental results are consistent with the exponentially fast convergence rate from theoretical analysis. Our result gives the first bound on the convergence rate of the co-occurrence matrix and the first sample complexity analysis in graph representation learning.