Goto

Collaborating Authors

 scalable inference


Scalable Inference in SDEs by Direct Matching of the Fokker-Planck-Kolmogorov Equation

Neural Information Processing Systems

This supplementary document is organized as follows. We provide details in terms of the concept of'solution' to an SDE, how we use a finite-differences As illustrated in Figure 1 in the main paper, the concept of a'solution' to an SDE is broader than that of This is what is done in this paper. We can now interpret Eq. (7) through these finite difference The model which we call a'GP-SDE' model in the main paper has appeared in various forms in literature before. It directly resembles a'random' ODE model, where the random field Figure 1 in the main paper, just providing further examples from the test set. For the timing experiments in Sec. 3, we constructed a setup that allowed us to control the approximation error.


Scalable inference of topic evolution via models for latent geometric structures

Neural Information Processing Systems

We develop new models and algorithms for learning the temporal dynamics of the topic polytopes and related geometric objects that arise in topic model based inference. Our model is nonparametric Bayesian and the corresponding inference algorithm is able to discover new topics as the time progresses. By exploiting the connection between the modeling of topic polytope evolution, Beta-Bernoulli process and the Hungarian matching algorithm, our method is shown to be several orders of magnitude faster than existing topic modeling approaches, as demonstrated by experiments working with several million documents in under two dozens of minutes.


Scalable Inference of Sparsely-changing Gaussian Markov Random Fields

Neural Information Processing Systems

We study the problem of inferring time-varying Gaussian Markov random fields, where the underlying graphical model is both sparse and changes {sparsely} over time. Most of the existing methods for the inference of time-varying Markov random fields (MRFs) rely on the \textit{regularized maximum likelihood estimation} (MLE), that typically suffer from weak statistical guarantees and high computational time. Instead, we introduce a new class of constrained optimization problems for the inference of sparsely-changing Gaussian MRFs (GMRFs). The proposed optimization problem is formulated based on the exact $\ell_0$ regularization, and can be solved in near-linear time and memory. Moreover, we show that the proposed estimator enjoys a provably small estimation error. We derive sharp statistical guarantees in the high-dimensional regime, showing that such problems can be learned with as few as one sample per time period. Our proposed method is extremely efficient in practice: it can accurately estimate sparsely-changing GMRFs with more than 500 million variables in less than one hour.


Scalable Inference in SDEs by Direct Matching of the Fokker–Planck–Kolmogorov Equation

Neural Information Processing Systems

Simulation-based techniques such as variants of stochastic Runge-Kutta are the de facto approach for inference with stochastic differential equations (SDEs) in machine learning. These methods are general-purpose and used with parametric and non-parametric models, and neural SDEs. Stochastic Runge-Kutta relies on the use of sampling schemes that can be inefficient in high dimensions. We address this issue by revisiting the classical SDE literature and derive direct approximations to the (typically intractable) Fokker-Planck-Kolmogorov equation by matching moments. We show how this workflow is fast, scales to high-dimensional latent spaces, and is applicable to scarce-data applications, where a non-parametric SDE with a driving Gaussian process velocity field specifies the model.



Scalable Inference for Logistic-Normal Topic Models

Neural Information Processing Systems

Logistic-normal topic models can effectively discover correlation structures among latent topics. However, their inference remains a challenge because of the non-conjugacy between the logistic-normal prior and multinomial topic mixing proportions. Existing algorithms either make restricting mean-field assumptions or are not scalable to large-scale applications. This paper presents a partially collapsed Gibbs sampling algorithm that approaches the provably correct distribution by exploring the ideas of data augmentation. To improve time efficiency, we further present a parallel implementation that can deal with large-scale applications and learn the correlation structures of thousands of topics from millions of documents. Extensive empirical results demonstrate the promise.


Scalable Inference for Neuronal Connectivity from Calcium Imaging

Neural Information Processing Systems

Fluorescent calcium imaging provides a potentially powerful tool for inferring connectivity in neural circuits with up to thousands of neurons. However, a key challenge in using calcium imaging for connectivity detection is that current systems often have a temporal response and frame rate that can be orders of magnitude slower than the underlying neural spiking process. Bayesian inference based on expectation-maximization (EM) have been proposed to overcome these limitations, but they are often computationally demanding since the E-step in the EM procedure typically involves state estimation in a high-dimensional nonlinear dynamical system. In this work, we propose a computationally fast method for the state estimation based on a hybrid of loopy belief propagation and approximate message passing (AMP). The key insight is that a neural system as viewed through calcium imaging can be factorized into simple scalar dynamical systems for each neuron with linear interconnections between the neurons. Using the structure, the updates in the proposed hybrid AMP methodology can be computed by a set of one-dimensional state estimation procedures and linear transforms with the connectivity matrix. This yields a computationally scalable method for inferring connectivity of large neural circuits. Simulations of the method on realistic neural networks demonstrate good accuracy with computation times that are potentially significantly faster than current approaches based on Markov Chain Monte Carlo methods.


Scalable Inference in SDEs by Direct Matching of the Fokker-Planck-Kolmogorov Equation

Neural Information Processing Systems

This supplementary document is organized as follows. We provide details in terms of the concept of'solution' to an SDE, how we use a finite-differences As illustrated in Figure 1 in the main paper, the concept of a'solution' to an SDE is broader than that of This is what is done in this paper. We can now interpret Eq. (7) through these finite difference The model which we call a'GP-SDE' model in the main paper has appeared in various forms in literature before. It directly resembles a'random' ODE model, where the random field Figure 1 in the main paper, just providing further examples from the test set. For the timing experiments in Sec. 3, we constructed a setup that allowed us to control the approximation error.


Scalable inference of topic evolution via models for latent geometric structures

Neural Information Processing Systems

We develop new models and algorithms for learning the temporal dynamics of the topic polytopes and related geometric objects that arise in topic model based inference. Our model is nonparametric Bayesian and the corresponding inference algorithm is able to discover new topics as the time progresses. By exploiting the connection between the modeling of topic polytope evolution, Beta-Bernoulli process and the Hungarian matching algorithm, our method is shown to be several orders of magnitude faster than existing topic modeling approaches, as demonstrated by experiments working with several million documents in under two dozens of minutes.


Reviews: Scalable inference of topic evolution via models for latent geometric structures

Neural Information Processing Systems

This is a very well written paper, both in style and substance. There are a few stylistic peculiarities that could surely be ruled out by thorough proof-reading. The authors present a nice introduction into the idea of modelling sets of topics, i.e. sets of points on a simplex, as the geometric structure of a polytope. They go on to describe, how evolution of such a polytope can be modelled over time by embedding a unit hypersphere into the simplex and modelling polytope evolution as random trajectories over this sphere. They further present a non-parametric hierarchical model for capturing polytopes with a varying number of topics and also multiple polytopes arising from different corpora.