Goto

Collaborating Authors

Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation

Neural Information Processing Systems

Recent work has shown that state-of-the-art classifiers are quite brittle, in the sense that a small adversarial change of an originally with high confidence correctly classified input leads to a wrong classification again with high confidence. This raises concerns that such classifiers are vulnerable to attacks and calls into question their usage in safety-critical systems. We show in this paper for the first time formal guarantees on the robustness of a classifier by giving instance-specific \emph{lower bounds} on the norm of the input manipulation required to change the classifier decision. Based on this analysis we propose the Cross-Lipschitz regularization functional. We show that using this form of regularization in kernel methods resp.


Learning Graph Representations with Embedding Propagation

Neural Information Processing Systems

We propose EP, Embedding Propagation, an unsupervised learning framework for graph-structured data. EP learns vector representations of graphs by passing two types of messages between neighboring nodes. Forward messages consist of label representations such as representations of words and other attributes associated with the nodes. Backward messages consist of gradients that result from aggregating the label representations and applying a reconstruction loss. Node representations are finally computed from the representation of their labels. With significantly fewer parameters and hyperparameters, an instance of EP is competitive with and often outperforms state of the art unsupervised and semi-supervised learning methods on a range of benchmark data sets.


Learning ReLUs via Gradient Descent

Neural Information Processing Systems

In this paper we study the problem of learning Rectified Linear Units (ReLUs) which are functions of the form $\vct{x}\mapsto \max(0,\langle \vct{w},\vct{x}\rangle)$ with $\vct{w}\in\R^d$ denoting the weight vector. We study this problem in the high-dimensional regime where the number of observations are fewer than the dimension of the weight vector. We assume that the weight vector belongs to some closed set (convex or nonconvex) which captures known side-information about its structure. We focus on the realizable model where the inputs are chosen i.i.d.~from a Gaussian distribution and the labels are generated according to a planted weight vector. We show that projected gradient descent, when initialized at $\vct{0}$, converges at a linear rate to the planted model with a number of samples that is optimal up to numerical constants. Our results on the dynamics of convergence of these very shallow neural nets may provide some insights towards understanding the dynamics of deeper architectures.


Bayesian Dyadic Trees and Histograms for Regression

Neural Information Processing Systems

Many machine learning tools for regression are based on recursive partitioning of the covariate space into smaller regions, where the regression function can be estimated locally. Among these, regression trees and their ensembles have demonstrated impressive empirical performance. In this work, we shed light on the machinery behind Bayesian variants of these methods. In particular, we study Bayesian regression histograms, such as Bayesian dyadic trees, in the simple regression case with just one predictor. We focus on the reconstruction of regression surfaces that are piecewise constant, where the number of jumps is unknown.


Convergence of Gradient EM on Multi-component Mixture of Gaussians

Neural Information Processing Systems

We derive the convergence rate depending on the mixing coefficients, minimum and maximum pairwise distances between the true centers, dimensionality and number of components; and obtain a near-optimal local contraction radius. While there have been some recent notable works that derive local convergence rates for EM in the two symmetric mixture of Gaussians, in the more general case, the derivations need structurally different and non-trivial arguments. We use recent tools from learning theory and empirical processes to achieve our theoretical results.


Starmer, Zelenskyy urge 'focus' on Ukraine as Iran war diverts attention

Al Jazeera

How the US left Ukraine exposed to Russia's winter war Will Europe use frozen Russian assets to fund war? How can Ukraine rebuild China ties? Starmer, Zelenskyy urge'focus' on Ukraine as Iran war diverts attention Ukrainian President Volodymyr Zelenskyy has met British Prime Minister Keir Starmer in London to sign a new defence pact as the unfolding US-Israeli war on Iran threatened to divert international attention away from Russia's attacks on Ukraine. Starmer welcomed Zelenskyy at his official Downing Street residence on Tuesday, reassuring the Ukrainian leader that "the focus must remain on Ukraine", days after the US partially rolled back sanctions against Moscow to cool oil prices sent soaring by its attacks on Iran. "There is obviously a conflict in Iran going on, in the Middle East, but we can't lose focus on what's going on in Ukraine and the need for our support there," said Starmer, who was meeting Zelenskyy to sign a defence partnership aimed at boosting "global defensive capability against the proliferation of low-cost, high-tech military hardware".


Google Intelligence is now asking for all of your data

PCWorld

Google is integrating its AI assistant Gemini with all personal Google services, allowing the AI to access user data across accounts for enhanced search results. PCWorld reports this opt-in feature enables Gemini to answer complex queries by collating personal information, though it's unavailable for business accounts. While offering improved functionality in Google AI Search and Chrome, this integration raises significant privacy concerns about Google's expanded data access. Google is now offering you the ability to link all of your Google services to the company's AI services for better results, looping in Google Intelligence, Gemini, with Gmail and your other data. It's a classic tactic used from Microsoft to Google and others: Connect your sources of data, and the service will become more adept at predicting what you want.


Differentially private Bayesian learning on distributed data

Neural Information Processing Systems

Many applications of machine learning, for example in health care, would benefit from methods that can guarantee privacy of data subjects. Differential privacy (DP) has become established as a standard for protecting learning results. The standard DP algorithms require a single trusted party to have access to the entire data, which is a clear weakness, or add prohibitive amounts of noise. We consider DP Bayesian learning in a distributed setting, where each party only holds a single sample or a few samples of the data. We propose a learning strategy based on a secure multi-party sum function for aggregating summaries from data holders and the Gaussian mechanism for DP. Our method builds on an asymptotically optimal and practically efficient DP Bayesian inference with rapidly diminishing extra cost.


Multi-Information Source Optimization

Neural Information Processing Systems

We consider Bayesian methods for multi-information source optimization (MISO), in which we seek to optimize an expensive-to-evaluate black-box objective function while also accessing cheaper but biased and noisy approximations (information sources). We present a novel algorithm that outperforms the state of the art for this problem by using a Gaussian process covariance kernel better suited to MISO than those used by previous approaches, and an acquisition function based on a one-step optimality analysis supported by efficient parallelization. We also provide a novel technique to guarantee the asymptotic quality of the solution provided by this algorithm. Experimental evaluations demonstrate that this algorithm consistently finds designs of higher value at less cost than previous approaches.


Practical Bayesian Optimization for Model Fitting with Bayesian Adaptive Direct Search

Neural Information Processing Systems

Computational models in fields such as computational neuroscience are often evaluated via stochastic simulation or numerical approximation. Fitting these models implies a difficult optimization problem over complex, possibly noisy parameter landscapes. Bayesian optimization (BO) has been successfully applied to solving expensive black-box problems in engineering and machine learning. Here we explore whether BO can be applied as a general tool for model fitting. First, we present a novel hybrid BO algorithm, Bayesian adaptive direct search (BADS), that achieves competitive performance with an affordable computational overhead for the running time of typical models. We then perform an extensive benchmark of BADS vs. many common and state-of-the-art nonconvex, derivative-free optimizers, on a set of model-fitting problems with real data and models from six studies in behavioral, cognitive, and computational neuroscience. With default settings, BADS consistently finds comparable or better solutions than other methods, including `vanilla' BO, showing great promise for advanced BO techniques, and BADS in particular, as a general model-fitting tool.