GibbsNet: Iterative Adversarial Inference for Deep Graphical Models

Lamb, Alex M., Hjelm, Devon, Ganin, Yaroslav, Cohen, Joseph Paul, Courville, Aaron C., Bengio, Yoshua

Neural Information Processing Systems 

Directed latent variable models that formulate the joint distribution as $p(x,z) p(z) p(x \mid z)$ have the advantage of fast and exact sampling. However, these models have the weakness of needing to specify $p(z)$, often with a simple fixed prior that limits the expressiveness of the model. Undirected latent variable models discard the requirement that $p(z)$ be specified with a prior, yet sampling from them generally requires an iterative procedure such as blocked Gibbs-sampling that may require many steps to draw samples from the joint distribution $p(x, z)$. We propose a novel approach to learning the joint distribution between the data and a latent code which uses an adversarially learned iterative procedure to gradually refine the joint distribution, $p(x, z)$, to better match with the data distribution on each step. GibbsNet is the best of both worlds both in theory and in practice.