Goto

Collaborating Authors

Hamiltonian Variational Auto-Encoder

Neural Information Processing Systems

Variational Auto-Encoders (VAE) have become very popular techniques to perform inference and learning in latent variable models as they allow us to leverage the rich representational power of neural networks to obtain flexible approximations of the posterior of latent variables as well as tight evidence lower bounds (ELBO). Combined with stochastic variational inference, this provides a methodology scaling to large datasets. However, for this methodology to be practically efficient, it is necessary to obtain low-variance unbiased estimators of the ELBO and its gradients with respect to the parameters of interest. While the use of Markov chain Monte Carlo (MCMC) techniques such as Hamiltonian Monte Carlo (HMC) has been previously suggested to achieve this [23, 26], the proposed methods require specifying reverse kernels which have a large impact on performance. Additionally, the resulting unbiased estimator of the ELBO for most MCMC kernels is typically not amenable to the reparameterization trick. We show here how to optimally select reverse kernels in this setting and, by building upon Hamiltonian Importance Sampling (HIS) [17], we obtain a scheme that provides low-variance unbiased estimators of the ELBO and its gradients using the reparameterization trick. This allows us to develop a Hamiltonian Variational Auto-Encoder (HVAE). This method can be re-interpreted as a target-informed normalizing flow [20] which, within our context, only requires a few evaluations of the gradient of the sampled likelihood and trivial Jacobian calculations at each iteration.


The best Kindles

Popular Science

Amazon's eReaders are best-in-class, and offer a legitimate opportunity for distraction-free reading. We may earn revenue from the products available on this page and participate in affiliate programs. The right Kindle will reignite your love of reading. Using a Kindle may seem unnecessary in a world where reading books, articles, and any other text on a phone or tablet is easy. Carrying around a dedicated mono-tasking device will add weight to your load, and it's another gadget to keep track of and charge. Yet Kindles remain popular because they only have one job and do it very well: let you carry and consume the stories that captivate you. A Kindle's e-ink screen won't reflect the sun when reading outdoors, unlike the reflective LCD displays used on phones and tablets.


Inferring Networks From Random Walk-Based Node Similarities

Neural Information Processing Systems

Digital presence in the world of online social media entails significant privacy risks. In this work we consider a privacy threat to a social network in which an attacker has access to a subset of random walk-based node similarities, such as effective resistances (i.e., commute times) or personalized PageRank scores. Using these similarities, the attacker seeks to infer as much information as possible about the network, including unknown pairwise node similarities and edges. For the effective resistance metric, we show that with just a small subset of measurements, one can learn a large fraction of edges in a social network. We also show that it is possible to learn a graph which accurately matches the underlying network on all other effective resistances.


Zach Braff shuts down rumors he has an AI chatbot girlfriend: 'Please update all gossip sites'

FOX News

Zach Braff AI chatbot girlfriend rumors denied by the actor on Instagram after podcast speculation. The Scrubs star clarified he is not dating an AI chatbot and called claims false.


Forward Modeling for Partial Observation Strategy Games - A StarCraft Defogger

Neural Information Processing Systems

We formulate the problem of defogging as state estimation and future state prediction from previous, partial observations in the context of real-time strategy games. We propose to employ encoder-decoder neural networks for this task, and introduce proxy tasks and baselines for evaluation to assess their ability of capturing basic game rules and high-level dynamics. By combining convolutional neural networks and recurrent networks, we exploit spatial and sequential correlations and train well-performing models on a large dataset of human games of StarCraft: Brood War. Finally, we demonstrate the relevance of our models to downstream tasks by applying them for enemy unit prediction in a state-of-the-art, rule-based StarCraft bot. We observe improvements in win rates against several strong community bots.


When do random forests fail?

Neural Information Processing Systems

Random forests are learning algorithms that build large collections of random trees and make predictions by averaging the individual tree predictions. In this paper, we consider various tree constructions and examine how the choice of parameters affects the generalization error of the resulting random forests as the sample size goes to infinity. We show that subsampling of data points during the tree construction phase is important: Forests can become inconsistent with either no subsampling or too severe subsampling. As a consequence, even highly randomized trees can lead to inconsistent forests if no subsampling is used, which implies that some of the commonly used setups for random forests can be inconsistent. As a second consequence we can show that trees that have good performance in nearest-neighbor search can be a poor choice for random forests.


7 Kindle settings you should change

Popular Science

Make sure your e-reader is set up exactly the way you want it. There are plenty of ways to tweak how your Kindle works. Breakthroughs, discoveries, and DIY tips sent six days a week. All of the Amazon Kindle models are intentionally designed to be straightforward to use. Grab your Kindle, tap the power button, and you're back reading from the place you left off (it's almost as simple as opening a real book).


The Spectrum of the Fisher Information Matrix of a Single-Hidden-Layer Neural Network

Neural Information Processing Systems

An important factor contributing to the success of deep learning has been the remarkable ability to optimize large neural networks using simple first-order optimization algorithms like stochastic gradient descent. While the efficiency of such methods depends crucially on the local curvature of the loss surface, very little is actually known about how this geometry depends on network architecture and hyperparameters. In this work, we extend a recently-developed framework for studying spectra of nonlinear random matrices to characterize an important measure of curvature, namely the eigenvalues of the Fisher information matrix. We focus on a single-hidden-layer neural network with Gaussian data and weights and provide an exact expression for the spectrum in the limit of infinite width. We find that linear networks suffer worse conditioning than nonlinear networks and that nonlinear networks are generically non-degenerate. We also predict and demonstrate empirically that by adjusting the nonlinearity, the spectrum can be tuned so as to improve the efficiency of first-order optimization methods.


SPIDER: Near-Optimal Non-Convex Optimization via Stochastic Path-Integrated Differential Estimator

Neural Information Processing Systems

In this paper, we propose a new technique named \textit{Stochastic Path-Integrated Differential EstimatoR} (SPIDER), which can be used to track many deterministic quantities of interests with significantly reduced computational cost. Combining SPIDER with the method of normalized gradient descent, we propose SPIDER-SFO that solve non-convex stochastic optimization problems using stochastic gradients only. We provide a few error-bound results on its convergence rates. Specially, we prove that the SPIDER-SFO algorithm achieves a gradient computation cost of $\mathcal{O}\left( \min( n^{1/2} \epsilon^{-2}, \epsilon^{-3}) \right)$ to find an $\epsilon$-approximate first-order stationary point. In addition, we prove that SPIDER-SFO nearly matches the algorithmic lower bound for finding stationary point under the gradient Lipschitz assumption in the finite-sum setting.


The world's oldest wild bird has a new grandchick

Popular Science

Environment Animals Wildlife Birds The world's oldest wild bird has a new grandchick Biologists have been tracking Wisdom, the roughly 75-year-old Laysan albatross, since the 1950s. Albatross chicks are getting stronger. Breakthroughs, discoveries, and DIY tips sent six days a week. The U.S. Fish and Wildlife Service is shining a light on a new member of a famous feathered family--that of the world's oldest known breeding bird, a Laysan albatross called Wisdom. The agency posted a video on social media featuring a scruffy looking hatchling seemingly yawning as it hangs out in the sand in close contact with a giant bird --presumably one of its parents.