vae
Gaussian Process Prior Variational Autoencoders
Variational autoencoders (VAE) are a powerful and widely-used class of models to learn complex data distributions in an unsupervised fashion. One important limitation of VAEs is the prior assumption that latent sample representations are independent and identically distributed. However, for many important datasets, such as time-series of images, this assumption is too strong: accounting for covariances between samples, such as those in time, can yield to a more appropriate model specification and improve performance in downstream tasks. In this work, we introduce a new model, the Gaussian Process (GP) Prior Variational Autoencoder (GPPVAE), to specifically address this issue. The GPPVAE aims to combine the power of VAEs with the ability to model correlations afforded by GP priors. To achieve efficient inference in this new class of models, we leverage structure in the covariance matrix, and introduce a new stochastic backpropagation strategy that allows for computing stochastic gradients in a distributed and low-memory fashion. We show that our method outperforms conditional VAEs (CVAEs) and an adaptation of standard VAEs in two image data applications.
- North America > United States > California (0.14)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
e04101138a3c94544760c1dbdf2c7a2d-Paper-Conference.pdf
For example, while prior work has suggested that theglobally optimal VAEsolution canlearn thecorrect manifold dimension, anecessary (butnotsufficient)condition forproducing samplesfrom the true data distribution, this has never been rigorously proven. Moreover, it remains unclear how such considerations would change when various types of conditioning variablesare introduced, or when the data support is extended to a union of manifolds (e.g., as is likely the case for MNIST digits and related). In this work, we address these points by first proving that VAE global minima are indeed capable of recovering the correct manifold dimension.
e3844e186e6eb8736e9f53c0c5889527-Paper.pdf
Inference networks oftraditional Variational Autoencoders (VAEs) aretypically amortized, resulting in relatively inaccurate posterior approximation compared to instance-wise variational optimization. Recent semi-amortized approaches were proposedtoaddress thisdrawback; however,theiriterativegradient update procedures can be computationally demanding.
- North America > United States > New Jersey > Middlesex County > Piscataway (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
ModelSelectionforBayesianAutoencoders: SupplementaryMaterial
In this section, we review some key results on the Wasserstein distance. Wpp Rπ(t,θi),Rρ(t,θi), (4) where the approximation comes from using Monte-Carlo integration by samplingθi uniformly in SD 1 [2]. M,M is the number of points used to approximate the integral. Calculating the Wasserstein distance with the empirical distribution function is computationally attractive. To do that, we first sortxms in an ascending order, such thatxi[m] xi[m+1], where i[m]istheindexofthesortedxms. Hamiltonian Monte Carlo (HMC)[24]isahighly-efficient MarkovChain Monte Carlo (MCMC) method used to generate samples from the posteriorw p(w|y).
- Europe > France (0.05)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- (8 more...)
ImprovingVariationalAutoencoderswithDensity Gap-based Regularization
On that basis, we hypothesize that these two problems stem from the conflict between the KL regularization inELBo andthefunction definition oftheprior distribution. Assuch, wepropose a novel regularization to substitute the KL regularization in ELBo for VAEs, which isbased on the density gapbetween the aggregated posterior distribution and the prior distribution.