Goto

Collaborating Authors

 Opper, Manfred


Inferring Parameter Distributions in Heterogeneous Motile Particle Ensembles: A Likelihood Approach for Second Order Langevin Models

arXiv.org Machine Learning

The inherent complexity of biological agents often leads to motility behavior that appears to have random components. Robust stochastic inference methods are therefore required to understand and predict the motion patterns from time discrete trajectory data provided by experiments. In many cases second order Langevin models are needed to adequately capture the motility. Additionally, population heterogeneity needs to be taken into account when analyzing data from several individual organisms. In this work, we describe a maximum likelihood approach to infer dynamical, stochastic models and, simultaneously, estimate the heterogeneity in a population of motile active particles from discretely sampled, stochastic trajectories. To this end we propose a new method to approximate the likelihood for non-linear second order Langevin models. We show that this maximum likelihood ansatz outperforms alternative approaches especially for short trajectories. Additionally, we demonstrate how a measure of uncertainty for the heterogeneity estimate can be derived. We thereby pave the way for the systematic, data-driven inference of dynamical models for actively driven entities based on trajectory data, deciphering temporal fluctuations and inter-particle variability.


Bridging discrete and continuous state spaces: Exploring the Ehrenfest process in time-continuous diffusion models

arXiv.org Machine Learning

While the original attempt considers processes can be discrete or continuous. In this discrete-time, continuous-space processes (Ho et al., work, we study time-continuous Markov jump 2020), one can show that in the small step-size limit the processes on discrete state spaces and investigate models converge to continuous-time, continuous-space processes their correspondence to state-continuous diffusion given by stochastic differential equations (SDEs) processes given by SDEs. In particular, we revisit (Song et al., 2021). This continuous time framework then the Ehrenfest process, which converges to an allows fruitful connections to mathematical tools such as Ornstein-Uhlenbeck process in the infinite state partial differential equations, path space measures and optimal space limit. Likewise, we can show that the timereversal control (Berner et al., 2024). As an alternative, one of the Ehrenfest process converges to the can consider discrete state spaces in continuous time via time-reversed Ornstein-Uhlenbeck process. This Markov jump processes, which have been suggested for observation bridges discrete and continuous state generative modeling in Campbell et al. (2022). Those are spaces and allows to carry over methods from one particularly promising for problems that naturally operate to the respective other setting. Additionally, we on discrete data, such as, e.g., text, images, graph structures suggest an algorithm for training the time-reversal or certain biological data, to name just a few.


A Convergence Analysis of Approximate Message Passing with Non-Separable Functions and Applications to Multi-Class Classification

arXiv.org Artificial Intelligence

Motivated by the recent application of approximate message passing (AMP) to the analysis of convex optimizations in multi-class classifications [Loureiro, et. al., 2021], we present a convergence analysis of AMP dynamics with non-separable multivariate nonlinearities. As an application, we present a complete (and independent) analysis of the motivated convex optimization problem.


Variational Inference for SDEs Driven by Fractional Noise

arXiv.org Machine Learning

We present a novel variational framework for performing inference in (neural) stochastic differential equations (SDEs) driven by Markov-approximate fractional Brownian motion (fBM). SDEs offer a versatile tool for modeling real-world continuous-time dynamic systems with inherent noise and randomness. Combining SDEs with the powerful inference capabilities of variational methods, enables the learning of representative function distributions through stochastic gradient descent. However, conventional SDEs typically assume the underlying noise to follow a Brownian motion (BM), which hinders their ability to capture long-term dependencies. In contrast, fractional Brownian motion (fBM) extends BM to encompass non-Markovian dynamics, but existing methods for inferring fBM parameters are either computationally demanding or statistically inefficient. In this paper, building upon the Markov approximation of fBM, we derive the evidence lower bound essential for efficient variational inference of posterior path measures, drawing from the well-established field of stochastic analysis. Additionally, we provide a closed-form expression to determine optimal approximation coefficients. Furthermore, we propose the use of neural networks to learn the drift, diffusion and control terms within our variational posterior, leading to the variational training of neural-SDEs. In this framework, we also optimize the Hurst index, governing the nature of our fractional noise. Beyond validation on synthetic data, we contribute a novel architecture for variational latent video prediction,-an approach that, to the best of our knowledge, enables the first variational neural-SDE application to video perception.


Analysis of Random Sequential Message Passing Algorithms for Approximate Inference

arXiv.org Artificial Intelligence

We analyze the dynamics of a random sequential message passing algorithm for approximate inference with large Gaussian latent variable models in a student-teacher scenario. To model nontrivial dependencies between the latent variables, we assume random covariance matrices drawn from rotation invariant ensembles. Moreover, we consider a model mismatching setting, where the teacher model and the one used by the student may be different. By means of dynamical functional approach, we obtain exact dynamical mean-field equations characterizing the dynamics of the inference algorithm. We also derive a range of model parameters for which the sequential algorithm does not converge. The boundary of this parameter range coincides with the de Almeida Thouless (AT) stability condition of the replica symmetric ansatz for the static probabilistic model.


Adaptive Inducing Points Selection For Gaussian Processes

arXiv.org Machine Learning

Gaussian Processes (\textbf{GPs}) are flexible non-parametric models with strong probabilistic interpretation. While being a standard choice for performing inference on time series, GPs have few techniques to work in a streaming setting. \cite{bui2017streaming} developed an efficient variational approach to train online GPs by using sparsity techniques: The whole set of observations is approximated by a smaller set of inducing points (\textbf{IPs}) and moved around with new data. Both the number and the locations of the IPs will affect greatly the performance of the algorithm. In addition to optimizing their locations, we propose to adaptively add new points, based on the properties of the GP and the structure of the data.


Nonlinear Hawkes Process with Gaussian Process Self Effects

arXiv.org Artificial Intelligence

Sequences of self exciting, or inhibiting, temporal events are frequent footmarks of natural phenomena: Earthquakes are known to be temporally clustered as aftershocks are commonly triggered following the occurrence of a main event [Ogata, 1988]; in social networks, the propagation of news can be modeled in terms of information cascades over the edges of a graph [Zhao et al., 2015]; and in neuronal activity, the occurrence of one spike may increase or decrease the probability of the occurrence of the next spike over some time period [Dayan and Abbott, 2001]. Traditionally, sequences of events in continuous time are modeled by Point processes, of which Cox processes [Cox, 1955], or doubly stochastic processes, use a stochastic process for the intensity function, which depends only on time and is not effected by the occurrences of the events. The Hawkes process [Hawkes and Oakes, 1974] extends the Cox process to capture phenomena in which the past events affects future arrivals, by introducing a memory dependence via a memory kernel. When incorporating dependence of the process on its own history, due to the superposition theorem of point process, new events will depend on either an exogenous rate, which is independent of the history, or an endogenous rate from past arrivals. This results in a branching structure, where new events that originate from previous events can be seen as "children" of the past events.


A Dynamical Mean-Field Theory for Learning in Restricted Boltzmann Machines

arXiv.org Machine Learning

We define a message-passing algorithm for computing magnetization s in Restricted Boltzmann machines, which are Ising models on bipartite g raphs introduced as neural network models for probability distributions over spin con figurations. To model nontrivial statistical dependencies between the spins' couplings, we assume that the rectangular coupling matrix is drawn from an arbitrary bi-rotation in variant random matrix ensemble. Using the dynamical functional method of statist ical mechanics we exactly analyze the dynamics of the algorithm in the large system limit. We prove the global convergence of the algorithm under a stability criterion and c ompute asymptotic convergence rates showing excellent agreement with numerical sim ulations.


Automated Augmented Conjugate Inference for Non-conjugate Gaussian Process Models

arXiv.org Machine Learning

We propose automated augmented conjugate inference, a new inference method for non-conjugate Gaussian processes (GP) models. Our method automatically constructs an auxiliary variable augmentation that renders the GP model conditionally conjugate. Building on the conjugate structure of the augmented model, we develop two inference methods. First, a fast and scalable stochastic variational inference method that uses efficient block coordinate ascent updates, which are computed in closed form. Second, an asymptotically correct Gibbs sampler that is useful for small datasets. Our experiments show that our method are up two orders of magnitude faster and more robust than existing state-of-the-art black-box methods.


Variational inference for Markov jump processes

Neural Information Processing Systems

Markov jump processes play an important role in a large number of application domains. However, realistic systems are analytically intractable and they have traditionally been analysed using simulation based techniques, which do not provide a framework for statistical inference. We propose a mean field approximation to perform posterior inference and parameter estimation. The approximation allows a practical solution to the inference problem, {while still retaining a good degree of accuracy.} We illustrate our approach on two biologically motivated systems.