Goto

Collaborating Authors

 Steidl, Gabriele


Generalized Normalizing Flows via Markov Chains

arXiv.org Artificial Intelligence

Normalizing flows, diffusion normalizing flows and variational autoencoders are powerful generative models. This chapter provides a unified framework to handle these approaches via Markov chains. We consider stochastic normalizing flows as a pair of Markov chains fulfilling some properties and show how many state-of-the-art models for data generation fit into this framework. Indeed numerical simulations show that including stochastic layers improves the expressivity of the network and allows for generating multimodal distributions from unimodal ones. The Markov chains point of view enables us to couple both deterministic layers as invertible neural networks and stochastic layers as Metropolis-Hasting layers, Langevin layers, variational autoencoders and diffusion normalizing flows in a mathematically sound way. Our framework establishes a useful mathematical tool to combine the various approaches.


Stochastic Normalizing Flows for Inverse Problems: a Markov Chains Viewpoint

arXiv.org Artificial Intelligence

Deep generative models for approximating complicated and often high-dimensional probability distributions became a rapidly developing research field. Normalizing flows are a popular subclass of these generative models. They can be used to model a target distribution by a simpler latent distribution which is usually the standard normal distribution. In this paper, we are interested in finite normalizing flows which are basically concatenations of learned diffeomorphisms. The parameters of the diffeomorphism are adapted to the target distribution by minimizing a loss functions. To this end, the diffeomorphism must have a tractable Jacobian determinant. For the continuous counterpart of normalizing flows, we refer to the overview paper [43] and the references therein. Suitable architectures of finite normalizing flows include invertible residual neural networks (ResNets) [7, 11, 22], (coupling-based) invertible neural networks (INNs) [4, 14, 29, 34, 40] and autoregessive flows [13, 15, 26, 38].


PCA Reduced Gaussian Mixture Models with Applications in Superresolution

arXiv.org Machine Learning

Despite the rapid development of computational hardware, the treatment of large and high dimensional data sets is still a challenging problem. This paper provides a twofold contribution to the topic. First, we propose a Gaussian Mixture Model in conjunction with a reduction of the dimensionality of the data in each component of the model by principal component analysis, called PCA-GMM. To learn the (low dimensional) parameters of the mixture model we propose an EM algorithm whose M-step requires the solution of constrained optimization problems. Fortunately, these constrained problems do not depend on the usually large number of samples and can be solved efficiently by an (inertial) proximal alternating linearized minimization algorithm. Second, we apply our PCA-GMM for the superresolution of 2D and 3D material images based on the approach of Sandeep and Jacob. Numerical results confirm the moderate influence of the dimensionality reduction on the overall superresolution result.


First order algorithms in variational image processing

arXiv.org Machine Learning

Variational methods in imaging are nowadays developing towards a quite universal and flexible tool, allowing for highly successful approaches on tasks like denoising, deblurring, inpainting, segmentation, super-resolution, disparity, and optical flow estimation. The overall structure of such approaches is of the form ${\cal D}(Ku) + \alpha {\cal R} (u) \rightarrow \min_u$ ; where the functional ${\cal D}$ is a data fidelity term also depending on some input data $f$ and measuring the deviation of $Ku$ from such and ${\cal R}$ is a regularization functional. Moreover $K$ is a (often linear) forward operator modeling the dependence of data on an underlying image, and $\alpha$ is a positive regularization parameter. While ${\cal D}$ is often smooth and (strictly) convex, the current practice almost exclusively uses nonsmooth regularization functionals. The majority of successful techniques is using nonsmooth and convex functionals like the total variation and generalizations thereof or $\ell_1$-norms of coefficients arising from scalar products with some frame system. The efficient solution of such variational problems in imaging demands for appropriate algorithms. Taking into account the specific structure as a sum of two very different terms to be minimized, splitting algorithms are a quite canonical choice. Consequently this field has revived the interest in techniques like operator splittings or augmented Lagrangians. Here we shall provide an overview of methods currently developed and recent results as well as some computational studies providing a comparison of different methods and also illustrating their success in applications.