Dubhashi, Devdatt
Stochastic Incremental Algorithms for Optimal Transport with SON Regularizer
Panahi, Ashkan, Thiel, Erik, Cheraghani, Morteza H., Dubhashi, Devdatt
We introduce a new regularizer for optimal transport (OT) which is tailored to better preserve the class structure. We give the first theoretical guarantees for an OT scheme that respects class structure. We give an accelerated proximal--projection scheme for this formulation with the proximal operator in closed form to give a highly scalable algorithm for computing optimal transport plans. We give a novel argument for the uniqueness of the optimum even in the absence of strong convexity. Our experiments show that the new regularizer preserves class structure better and is more robust compared to previous regularizers.
GANs for LIFE: Generative Adversarial Networks for Likelihood Free Inference
Jethava, Vinay, Dubhashi, Devdatt
We introduce a framework using Generative Adversarial Networks (GANs) for likelihood--free inference (LFI) and Approximate Bayesian Computation (ABC). Our approach addresses both the key problems in likelihood--free inference, namely how to compare distributions and how to efficiently explore the parameter space. Our framework allows one to use the simulator model as a black box and leverage the power of deep networks to generate a rich set of features in a data driven fashion (as opposed to previous ad hoc approaches). Thereby it is a step towards a powerful alternative approach to LFI and ABC. On benchmark data sets, our approach improves on others with respect to scalability, ability to handle high dimensional data and complex probability distributions.
Thompson Sampling for Stochastic Bandits with Graph Feedback
Tossou, Aristide C. Y. (Chalmers University of Technology) | Dimitrakakis, Christos (University of Lille, and Chalmers University of Technology) | Dubhashi, Devdatt (Chalmers University of Technology)
We present a simple set of algorithms based on Thompson Sampling for stochastic bandit problems with graph feedback. Thompson Sampling is generally applicable, without the need to construct complicated upper confidence bounds. As we show in this paper, it has excellent performance in problems with graph feedback, even when the graph structure itself is unknown and/or changing. We provide theoretical guarantees on the Bayesian regret of the algorithm, as well as extensive experi- mental results on real and simulated networks. More specifically, we tested our algorithms on power law, planted partitions and Erdo's–Rényi graphs, as well as on graphs derived from Facebook and Flixster data and show that they clearly outperform related methods that employ upper confidence bounds.
Weighted Theta Functions and Embeddings with Applications to Max-Cut, Clustering and Summarization
Johansson, Fredrik D., Chattoraj, Ankani, Bhattacharyya, Chiranjib, Dubhashi, Devdatt
We introduce a unifying generalization of the Lovász theta function, and the associated geometric embedding, for graphs with weights on both nodes and edges. We show how it can be computed exactly by semidefinite programming, and how to approximate it using SVM computations. We show how the theta function can be interpreted as a measure of diversity in graphs and use this idea, and the graph embedding in algorithms for Max-Cut, correlation clustering and document summarization, all of which are well represented as problems on weighted graphs.
The Lovász ϑ function, SVMs and finding large dense subgraphs
Jethava, Vinay, Martinsson, Anders, Bhattacharyya, Chiranjib, Dubhashi, Devdatt
The Lovasz $\theta$ function of a graph, is a fundamental tool in combinatorial optimization and approximation algorithms. Computing $\theta$ involves solving a SDP and is extremely expensive even for moderately sized graphs. In this paper we establish that the Lovasz $\theta$ function is equivalent to a kernel learning problem related to one class SVM. This interesting connection opens up many opportunities bridging graph theoretic algorithms and machine learning. We show that there exist graphs, which we call $SVM-\theta$ graphs, on which the Lovasz $\theta$ function can be approximated well by a one-class SVM. This leads to a novel use of SVM techniques to solve algorithmic problems in large graphs e.g. identifying a planted clique of size $\Theta({\sqrt{n}})$ in a random graph $G(n,\frac{1}{2})$. A classic approach for this problem involves computing the $\theta$ function, however it is not scalable due to SDP computation. We show that the random graph with a planted clique is an example of $SVM-\theta$ graph, and as a consequence a SVM based approach easily identifies the clique in large graphs and is competitive with the state-of-the-art. Further, we introduce the notion of a ''common orthogonal labeling'' which extends the notion of a ''orthogonal labelling of a single graph (used in defining the $\theta$ function) to multiple graphs. The problem of finding the optimal common orthogonal labelling is cast as a Multiple Kernel Learning problem and is used to identify a large common dense region in multiple graphs. The proposed algorithm achieves an order of magnitude scalability compared to the state of the art.