Information Constraints on Auto-Encoding Variational Bayes
Romain Lopez, Jeffrey Regier, Michael I. Jordan, Nir Yosef
–Neural Information Processing Systems
Parameterizing the approximate posterior of a generative model with neural networks has become a common theme in recent machine learning research. While providing appealing flexibility, this approach makes it difficult to impose or assess structural constraints such as conditional independence. We propose a framework for learning representations that relies on auto-encoding variational Bayes, in which the search space is constrained via kernel-based measures of independence. In particular, our method employs the d-variable Hilbert-Schmidt Independence Criterion (dHSIC) to enforce independence between the latent representations and arbitrary nuisance factors. We show how this method can be applied to a range of problems, including problems that involve learning invariant and conditionally independent representations. We also present a full-fledged application to singlecell RNA sequencing (scRNA-seq). In this setting the biological signal is mixed in complex ways with sequencing errors and sampling effects. We show that our method outperforms the state-of-the-art approach in this domain.
Neural Information Processing Systems
Oct-7-2024, 21:14:23 GMT
- Country:
- Asia (0.28)
- North America (0.46)
- Genre:
- Research Report > Promising Solution (0.66)
- Industry:
- Technology: