Goto

Collaborating Authors

 numpyro




Hamiltonian Monte Carlo Inference of Marginalized Linear Mixed-Effects Models

Lai, Jinlin, Domke, Justin, Sheldon, Daniel

arXiv.org Machine Learning

Bayesian reasoning in linear mixed-effects models (LMMs) is challenging and often requires advanced sampling techniques like Markov chain Monte Carlo (MCMC). A common approach is to write the model in a probabilistic programming language and then sample via Hamiltonian Monte Carlo (HMC). However, there are many ways a user can transform a model that make inference more or less efficient. In particular, marginalizing some variables can greatly improve inference but is difficult for users to do manually. We develop an algorithm to easily marginalize random effects in LMMs. A naive approach introduces cubic time operations within an inference algorithm like HMC, but we reduce the running time to linear using fast linear algebra techniques. We show that marginalization is always beneficial when applicable and highlight improvements in various models, especially ones from cognitive sciences.


BayesLDM: A Domain-Specific Language for Probabilistic Modeling of Longitudinal Data

Tung, Karine, De La Torre, Steven, Mistiri, Mohamed El, De Braganca, Rebecca Braga, Hekler, Eric, Pavel, Misha, Rivera, Daniel, Klasnja, Pedja, Spruijt-Metz, Donna, Marlin, Benjamin M.

arXiv.org Artificial Intelligence

In this paper we present BayesLDM, a system for Bayesian longitudinal data modeling consisting of a high-level modeling language with specific features for modeling complex multivariate time series data coupled with a compiler that can produce optimized probabilistic program code for performing inference in the specified model. BayesLDM supports modeling of Bayesian network models with a specific focus on the efficient, declarative specification of dynamic Bayesian Networks (DBNs). The BayesLDM compiler combines a model specification with inspection of available data and outputs code for performing Bayesian inference for unknown model parameters while simultaneously handling missing data. These capabilities have the potential to significantly accelerate iterative modeling workflows in domains that involve the analysis of complex longitudinal data by abstracting away the process of producing computationally efficient probabilistic inference code. We describe the BayesLDM system components, evaluate the efficiency of representation and inference optimizations and provide an illustrative example of the application of the system to analyzing heterogeneous and partially observed mobile health data.


An astronomer's introduction to NumPyro

#artificialintelligence

Over the past year or so, I've been using JAX extensively for my research, and I've also been encouraging other astronomers to give it a try. In particular, I've been using JAX as the computation engine for probabilistic inference tasks. There's more to it, but one way that I like to think about JAX is as NumPy with just-in-time compilation and automatic differentiation. The just-in-time compilation features of JAX can be used to speed up you NumPy computations by removing some Python overhead and by executing it on your GPU. Then, automatic differentiation can be used to efficiently compute the derivatives of your code with respect to its input parameters.


d3p -- A Python Package for Differentially-Private Probabilistic Programming

Prediger, Lukas, Loppi, Niki, Kaski, Samuel, Honkela, Antti

arXiv.org Machine Learning

Probabilistic modelling presents a natural way to model data by describing their (assumed) generative process. The model is then fit to observations by probabilistic inference algorithms. Learning from sensitive data, however, clearly raises concerns about privacy, calling for privacy-preserving model inference algorithms. Differential privacy (DP) [10] provides a rigorous mathematical framework for addressing such concerns and has become the de-facto standard notion for privacy. It essentially assures that an algorithms outputs will not differ significantly whether a specific individual's data record is included in the data set or not.


Composable Effects for Flexible and Accelerated Probabilistic Programming in NumPyro

Phan, Du, Pradhan, Neeraj, Jankowiak, Martin

arXiv.org Artificial Intelligence

NumPyro is a lightweight library that provides an alternate NumPy backend to the Pyro probabilistic programming language with the same modeling interface, language primitives and effect handling abstractions. Effect handlers allow Pyro's modeling API to be extended to NumPyro despite its being built atop a fundamentally different JAX-based functional backend. In this work, we demonstrate the power of composing Pyro's effect handlers with the program transformations that enable hardware acceleration, automatic differentiation, and vectorization in JAX. In particular, NumPyro provides an iterative formulation of the No-U-Turn Sampler (NUTS) that can be end-to-end JIT compiled, yielding an implementation that is much faster than existing alternatives in both the small and large dataset regimes.