Goto

Collaborating Authors

 peek search





1 3 5 7 9 11 13 15 17 19 46 45 44 43 42 41 40 39 Latency L Log-probability (scale 10000) Log-probability

Neural Information Processing Systems

We are grateful to the reviewers for their feedback. We address all their comments below. We will fix the typo on line 142. Reviewer #2: Thank you very much for your feedback. We provide additional results to address all your concerns.


Reviews: Online Markov Decoding: Lower Bounds and Near-Optimal Approximation Algorithms

Neural Information Processing Systems

The authors propose three online inference methods, i.e., peek search, randomized peek search, and peek reset, for Markov chain models. They also use a proof framework to prove the competitive ratio for each method. The proof method first computes the lower bound of ON, so that gets the relationship between the ON and OPT. And then it uses the properties of geometric series to find an upper bound of competitive ratio. The authors also design a dynamic programming method to efficiently compute reward accumulation, and prove its complexity.


Peek Search: Near-Optimal Online Markov Decoding

Garg, Vikas K., Pichkhadze, Tamar

arXiv.org Machine Learning

We resolve the fundamental problem of online decoding with ergodic Markov models. Specifically, we provide deterministic and randomized algorithms that are provably near-optimal under latency constraints with respect to the unconstrained offline optimal algorithm. Our algorithms admit efficient implementation via dynamic programs, and extend to (possibly adversarial) non-stationary or time-varying Markov settings as well. Moreover, we establish lower bounds in both deterministic and randomized settings subject to latency requirements, and prove that no online algorithm can perform significantly better than our algorithms.