Goto

Collaborating Authors

 training point


An efficient, accurate, and interpretable machine learning method for computing probability of failure

Zhu, Jacob, Estep, Donald

arXiv.org Machine Learning

We introduce a novel machine learning method called the Penalized Profile Support Vector Machine based on the Gabriel edited set for the computation of the probability of failure for a complex system as determined by a threshold condition on a computer model of system behavior. The method is designed to minimize the number of evaluations of the computer model while preserving the geometry of the decision boundary that determines the probability. It employs an adaptive sampling strategy designed to strategically allocate points near the boundary determining failure and builds a locally linear surrogate boundary that remains consistent with its geometry by strategic clustering of training points. We prove two convergence results and we compare the performance of the method against a number of state of the art classification methods on four test problems. We also apply the method to determine the probability of survival using the Lotka--Volterra model for competing species.



Representer Point Selection for Explaining Deep Neural Networks

Neural Information Processing Systems

We propose to explain the predictions of a deep neural network, by pointing to the set of what we call representer points in the training set, for a given test point prediction. Specifically, we show that we can decompose the pre-activation prediction of a neural network into a linear combination of activations of training points, with the weights corresponding to what we call representer values, which thus capture the importance of that training point on the learned parameters of the network. But it provides a deeper understanding of the network than simply training point influence: with positive representer values corresponding to excitatory training points, and negative values corresponding to inhibitory points, which as we show provides considerably more insight. Our method is also much more scalable, allowing for real-time feedback in a manner not feasible with influence functions.


Supervised learning pays attention

Craig, Erin, Tibshirani, Robert

arXiv.org Machine Learning

In-context learning with attention enables large neural networks to make context-specific predictions by selectively focusing on relevant examples. Here, we adapt this idea to supervised learning procedures such as lasso regression and gradient boosting, for tabular data. Our goals are to (1) flexibly fit personalized models for each prediction point and (2) retain model simplicity and interpretability. Our method fits a local model for each test observation by weighting the training data according to attention, a supervised similarity measure that emphasizes features and interactions that are predictive of the outcome. Attention weighting allows the method to adapt to heterogeneous data in a data-driven way, without requiring cluster or similarity pre-specification. Further, our approach is uniquely interpretable: for each test observation, we identify which features are most predictive and which training observations are most relevant. We then show how to use attention weighting for time series and spatial data, and we present a method for adapting pretrained tree-based models to distributional shift using attention-weighted residual corrections. Across real and simulated datasets, attention weighting improves predictive performance while preserving interpretability, and theory shows that attention-weighting linear models attain lower mean squared error than the standard linear model under mixture-of-models data-generating processes with known subgroup structure.


When unlearning is free: leveraging low influence points to reduce computational costs

Kleiman, Anat, Fisher, Robert, Deaner, Ben, Wieder, Udi

arXiv.org Artificial Intelligence

As concerns around data privacy in machine learning grow, the ability to unlearn, or remove, specific data points from trained models becomes increasingly important. While state of the art unlearning methods have emerged in response, they typically treat all points in the forget set equally. In this work, we challenge this approach by asking whether points that have a negligible impact on the model's learning need to be removed. Through a comparative analysis of influence functions across language and vision tasks, we identify subsets of training data with negligible impact on model outputs. Leveraging this insight, we propose an efficient unlearning framework that reduces the size of datasets before unlearning leading to significant computational savings (up to approximately 50 percent) on real world empirical examples.


An ensemble diversity approach to supervised binary hashing

Miguel A. Carreira-Perpinan, Ramin Raziperchikolaei

Neural Information Processing Systems

Binary hashing is a well-known approach for fast approximat e nearest-neighbor search in information retrieval. Much work has focused on af finity-based objective functions involving the hash functions or binary codes. The se objective functions encode neighborhood information between data points and ar e often inspired by manifold learning algorithms. They ensure that the hash fun ctions differ from each other through constraints or penalty terms that encourage c odes to be orthogonal or dissimilar across bits, but this couples the binary varia bles and complicates the already difficult optimization. W e propose a much simpler ap proach: we train each hash function (or bit) independently from each other, b ut introduce diversity among them using techniques from classifier ensembles. Surp risingly, we find that not only is this faster and trivially parallelizable, b ut it also improves over the more complex, coupled objective function, and achieves sta te-of-the-art precision and recall in experiments with image retrieval.



Representer Point Selection for Explaining Deep Neural Networks

Neural Information Processing Systems

We propose to explain the predictions of a deep neural network, by pointing to the set of what we call representer points in the training set, for a given test point prediction. Specifically, we show that we can decompose the pre-activation prediction of a neural network into a linear combination of activations of training points, with the weights corresponding to what we call representer values, which thus capture the importance of that training point on the learned parameters of the network. But it provides a deeper understanding of the network than simply training point influence: with positive representer values corresponding to excitatory training points, and negative values corresponding to inhibitory points, which as we show provides considerably more insight. Our method is also much more scalable, allowing for real-time feedback in a manner not feasible with influence functions.