Goto

Collaborating Authors

 sparse


Select-and-Sample for Spike-and-Slab Sparse Coding

Neural Information Processing Systems

Probabilistic inference serves as a popular model for neural processing. It is still unclear, however, how approximate probabilistic inference can be accurate and scalable to very high-dimensional continuous latent spaces. Especially as typical posteriors for sensory data can be expected to exhibit complex latent dependencies including multiple modes. Here, we study an approach that can efficiently be scaled while maintaining a richly structured posterior approximation under these conditions. As example model we use spike-and-slab sparse coding for V1 processing, and combine latent subspace selection with Gibbs sampling (select-and-sample).


Proximal SCOPE for Distributed Sparse Learning

Neural Information Processing Systems

Distributed sparse learning with a cluster of multiple machines has attracted much attention in machine learning, especially for large-scale applications with high-dimensional data. One popular way to implement sparse learning is to use L1 regularization. In this paper, we propose a novel method, called proximal SCOPE (pSCOPE), for distributed sparse learning with L1 regularization.



80f2f15983422987ea30d77bb531be86-Paper.pdf

Neural Information Processing Systems

Wethenseparate theoptimization process into two steps, corresponding to weight update and structure parameter update. For the former step, we use the conventional chain rule, which can be sparse via exploiting the sparse structure.





Nonparametric Density Estimation & Convergence Rates for GANs under Besov IPM Losses

Ananya Uppal, Shashank Singh, Barnabas Poczos

Neural Information Processing Systems

Along line ofwork has established convergence rates ofthe empirical distribution tothe true distribution in spaces as general as unbounded metric spaces [54, 25, 45]). In the Euclidean setting, this is well understood [14,2,18], although, to the best of our knowledge, minimax lower bounds have been proven only recently [45]; this setting intersects with our work in the caseσd = 1,σg = 0, pd =,matchingourminimaxrateofn 1/D+n 1/2.