Goto

Collaborating Authors

 original network


a1ada9947e0d683b4625f94c74104d73-Paper.pdf

Neural Information Processing Systems

Many networks in the brain are sparsely connected, and the brain eliminates synapses during development and learning. How could the brain decide which synapses to prune?





Learning the Number of Neurons in Deep Networks

Jose M. Alvarez, Mathieu Salzmann

Neural Information Processing Systems

Nowadays, the number of layers and of neurons in each layer of a deep network are typically set manually. While very deep and wide networks have proven effective in general, they come at a high memory and computation cost, thus making them impractical for constrained platforms.





empirical results and in our plots will now show network dynamics as well as spectra

Neural Information Processing Systems

We have empirically characterized noise-prune's performance on non-symmetric clustered networks (i.e., going beyond We will include an expanded set of these results in the manuscript. We have not yet characterized noise-prune's performance against these Building off of the Reviewer's language learning example, even here the dynamical patterns are We will now include some text on other benefits of sparsity in the Discussion. Blue shows equivalent curve for weight-based pruning. Noise-prune performs significantly better than pruning by weights.


Explaining, Fast and Slow: Abstraction and Refinement of Provable Explanations

Bassan, Shahaf, Elboher, Yizhak Yisrael, Ladner, Tobias, Althoff, Matthias, Katz, Guy

arXiv.org Artificial Intelligence

Despite significant advancements in post-hoc explainability techniques for neural networks, many current methods rely on heuristics and do not provide formally provable guarantees over the explanations provided. Recent work has shown that it is possible to obtain explanations with formal guarantees by identifying subsets of input features that are sufficient to determine that predictions remain unchanged using neural network verification techniques. Despite the appeal of these explanations, their computation faces significant scalability challenges. In this work, we address this gap by proposing a novel abstraction-refinement technique for efficiently computing provably sufficient explanations of neural network predictions. Our method abstracts the original large neural network by constructing a substantially reduced network, where a sufficient explanation of the reduced network is also provably sufficient for the original network, hence significantly speeding up the verification process. If the explanation is in sufficient on the reduced network, we iteratively refine the network size by gradually increasing it until convergence. Our experiments demonstrate that our approach enhances the efficiency of obtaining provably sufficient explanations for neural network predictions while additionally providing a fine-grained interpretation of the network's predictions across different abstraction levels.