Goto

Collaborating Authors

 efficient loss-based decoding


Efficient Loss-Based Decoding on Graphs for Extreme Classification

Neural Information Processing Systems

In extreme classification problems, learning algorithms are required to map instances to labels from an extremely large label set. We build on a recent extreme classification framework with logarithmic time and space (LTLS), and on a general approach for error correcting output coding (ECOC) with loss-based decoding, and introduce a flexible and efficient approach accompanied by theoretical bounds. Our framework employs output codes induced by graphs, for which we show how to perform efficient loss-based decoding to potentially improve accuracy. In addition, our framework offers a tradeoff between accuracy, model size and prediction time. We show how to find the sweet spot of this tradeoff using only the training data. Our experimental study demonstrates the validity of our assumptions and claims, and shows that our method is competitive with state-of-the-art algorithms.


Reviews: Efficient Loss-Based Decoding on Graphs for Extreme Classification

Neural Information Processing Systems

This paper proposes an algorithm to solve extreme multi-class classification problems using Error Correcting Output Coding. During training, the algorithm simply learns l (logK) independent binary classifiers. Main contribution of the paper is in the inference algorithm. It reduces the costly loss based decoding framework for ECOC to that of finding shortest path on a weighted trellis graph. The paper is well written and easy to understand.


Efficient Loss-Based Decoding On Graphs For Extreme Classification

Evron, Itay, Moroshko, Edward, Crammer, Koby

arXiv.org Machine Learning

In extreme classification problems, learning algorithms are required to map instances to labels from an extremely large label set. We build on a recent extreme classification framework with logarithmic time and space, and on a general approach for error correcting output coding (ECOC), and introduce a flexible and efficient approach accompanied by bounds. Our framework employs output codes induced by graphs, and offers a tradeoff between accuracy and model size. We show how to find the sweet spot of this tradeoff using only the training data. Our experimental study demonstrates the validity of our assumptions and claims, and shows the superiority of our method compared with state-of-the-art algorithms.