Not enough data to create a plot.
Try a different view from the menu above.
Country
Data Imputation through the Identification of Local Anomalies
Ozkan, Huseyin, Pelvan, Ozgun S., Kozat, Suleyman S.
We introduce a comprehensive and statistical framework in a model free setting for a complete treatment of localized data corruptions due to severe noise sources, e.g., an occluder in the case of a visual recording. Within this framework, we propose i) a novel algorithm to efficiently separate, i.e., detect and localize, possible corruptions from a given suspicious data instance and ii) a Maximum A Posteriori (MAP) estimator to impute the corrupted data. As a generalization to Euclidean distance, we also propose a novel distance measure, which is based on the ranked deviations among the data attributes and empirically shown to be superior in separating the corruptions. Our algorithm first splits the suspicious instance into parts through a binary partitioning tree in the space of data attributes and iteratively tests those parts to detect local anomalies using the nominal statistics extracted from an uncorrupted (clean) reference data set. Once each part is labeled as anomalous vs normal, the corresponding binary patterns over this tree that characterize corruptions are identified and the affected attributes are imputed. Under a certain conditional independency structure assumed for the binary patterns, we analytically show that the false alarm rate of the introduced algorithm in detecting the corruptions is independent of the data and can be directly set without any parameter tuning. The proposed framework is tested over several well-known machine learning data sets with synthetically generated corruptions; and experimentally shown to produce remarkable improvements in terms of classification purposes with strong corruption separation capabilities. Our experiments also indicate that the proposed algorithms outperform the typical approaches and are robust to varying training phase conditions.
Robustness and Generalization for Metric Learning
Bellet, Aurรฉlien, Habrard, Amaury
Metric learning has attracted a lot of interest over the last decade, but the generalization ability of such methods has not been thoroughly studied. In this paper, we introduce an adaptation of the notion of algorithmic robustness (previously introduced by Xu and Mannor) that can be used to derive generalization bounds for metric learning. We further show that a weak notion of robustness is in fact a necessary and sufficient condition for a metric learning algorithm to generalize. To illustrate the applicability of the proposed framework, we derive generalization results for a large family of existing metric learning algorithms, including some sparse formulations that are not covered by previous results.
Random forests with random projections of the output space for high dimensional multi-label classification
Joly, Arnaud, Geurts, Pierre, Wehenkel, Louis
We adapt the idea of random projections applied to the output space, so as to enhance tree-based ensemble methods in the context of multi-label classification. We show how learning time complexity can be reduced without affecting computational complexity and accuracy of predictions. We also show that random output space projections may be used in order to reach different bias-variance tradeoffs, over a broad panel of benchmark problems, and that this may lead to improved accuracy while reducing significantly the computational burden of the learning stage.
Distributed Variational Inference in Sparse Gaussian Process Regression and Latent Variable Models
Gal, Yarin, van der Wilk, Mark, Rasmussen, Carl E.
Gaussian processes (GPs) are a powerful tool for probabilistic inference over functions. They have been applied to both regression and non-linear dimensionality reduction, and offer desirable properties such as uncertainty estimates, robustness to over-fitting, and principled ways for tuning hyper-parameters. However the scalability of these models to big datasets remains an active topic of research. We introduce a novel re-parametrisation of variational inference for sparse GP regression and latent variable models that allows for an efficient distributed algorithm. This is done by exploiting the decoupling of the data given the inducing points to re-formulate the evidence lower bound in a Map-Reduce setting. We show that the inference scales well with data and computational resources, while preserving a balanced distribution of the load among the nodes. We further demonstrate the utility in scaling Gaussian processes to big data. We show that GP performance improves with increasing amounts of data in regression (on flight data with 2 million records) and latent variable modelling (on MNIST). The results show that GPs perform better than many common models often used for big data.
A Bayesian Tensor Factorization Model via Variational Inference for Link Prediction
Ermis, Beyza, Cemgil, A. Taylan
Probabilistic approaches for tensor factorization aim to extract meaningful structure from incomplete data by postulating low rank constraints. Recently, variational Bayesian (VB) inference techniques have successfully been applied to large scale models. This paper presents full Bayesian inference via VB on both single and coupled tensor factorization models. Our method can be run even for very large models and is easily implemented. It exhibits better prediction performance than existing approaches based on maximum likelihood on several real-world datasets for missing link prediction problem.
Communication-Efficient Distributed Dual Coordinate Ascent
Jaggi, Martin, Smith, Virginia, Takรกฤ, Martin, Terhorst, Jonathan, Krishnan, Sanjay, Hofmann, Thomas, Jordan, Michael I.
Communication remains the most significant bottleneck in the performance of distributed optimization algorithms for large-scale machine learning. In this paper, we propose a communication-efficient framework, CoCoA, that uses local computation in a primal-dual setting to dramatically reduce the amount of necessary communication. We provide a strong convergence rate analysis for this class of algorithms, as well as experiments on real-world distributed datasets with implementations in Spark. In our experiments, we find that as compared to state-of-the-art mini-batch versions of SGD and SDCA algorithms, CoCoA converges to the same .001-accurate solution quality on average 25x as quickly.
Variational Inference in Sparse Gaussian Process Regression and Latent Variable Models - a Gentle Tutorial
Gal, Yarin, van der Wilk, Mark
In this tutorial we explain the inference procedures developed for the sparse Gaussian process (GP) regression and Gaussian process latent variable model (GPLVM). Due to page limit the derivation given in Titsias (2009) and Titsias & Lawrence (2010) is brief, hence getting a full picture of it requires collecting results from several different sources and a substantial amount of algebra to fill-in the gaps. Our main goal is thus to collect all the results and full derivations into one place to help speed up understanding this work. In doing so we present a re-parametrisation of the inference that allows it to be carried out in parallel. A secondary goal for this document is, therefore, to accompany our paper and open-source implementation of the parallel inference scheme for the models. We hope that this document will bridge the gap between the equations as implemented in code and those published in the original papers, in order to make it easier to extend existing work. We assume prior knowledge of Gaussian processes and variational inference, but we also include references for further reading where appropriate.
Leveraging AI Teaching in the Cloud for AI Teaching on Campus
Fisher, Douglas H. (Vanderbilt University)
The Educational Advances in Artificial Intelligence column discusses and shares innovative educational approaches that teach or leverage AI and its many subfields at all levels of education (K-12, undergraduate, and graduate levels). I credit these positive changes to the active in-class learning and a new enthusiasm for teaching, as well as the first-rate lectures by Stanford professors Jennifer Wisdom and Andrew Ng. I was showed that students liked this SPOC format, although pleased when students, enrolled in Introduction to there were suggestions for better in-class and Artificial Intelligence Class MOOC CS188x at the MOOC-content coordination. Had I tweaked my University of California, Berkeley, came to my channel course and continued along this path, I might have for remediation, taking word back to the MOOC's achieved phenominal success, but sadly I left the discussion forum. I required students in my graduate SPOC format behind.
AAAI Conferences Calendar
This page includes forthcoming AAAI sponsored conferences, conferences presented by AAAI Affiliates, and conferences held in cooperation with AAAI. AI Magazine also maintains a calendar listing that includes nonaffiliated conferences at www.aaai.org/Magazine/calendar.php. AIIDE-14 will be held FLAIRS-15 will be held May 18-20, 10th ACM/IEEE International Conference October 3-7 in Raleigh, NC, USA 2015 in Hollywood, Florida, USA on Human-Robot Interaction. ICAART 2014 will be held January 10-12 in Lisbon, Portugal International Joint Conference on AAAI Fall Symposium Series. ICCBR 2014 held January 10-12 in Lisbon, Portugal will be held September 29 - October 1 AAAI Spring Symposium.
Computational Sustainability: Editorial Introduction to the Summer and Fall Issues
Eaton, Eric (University of Pennsylvania) | Gomes, Carla (Cornell University) | Williams, Brian C. (Massachusetts Institute of Technology)
Computational sustainability problems, which exist in dynamic environments with high amounts of uncertainty, provide a variety of unique challenges to artificial intelligence research and the opportunity for significant impact upon our collective future. This editorial introduction provides an overview of artificial intelligence for computational sustainability, and introduces the special issue articles that appear in this issue and the previous issue of AI Magazine.