Unsupervised or Indirectly Supervised Learning
Artificial intelligence can be harnessed to help older adults, others
The progress of technology in leaps and bounds has resulted in the generation of an enormous amount of digital data in the modern era. Against this backdrop, artificial intelligence (AI) has emerged as a useful mechanism to automatically organize and categorize data and to leverage useful patterns in the data to make intelligent predictions for the future from past observations. In this way, we can use AI to help older adults and others. The most common kind of AI algorithms is supervised machine learning, which involves learning from labeled data. However, while gathering a large amount of unlabeled data is cheap and easy, hand-labeling the data is an expensive process in terms of time, labor and human expertise.
Unsupervised learning demystified โ Cassie Kozyrkov โ Medium
Unsupervised learning sounds like a fancy way to say "let the kids learn on their own not to touch the hot oven", but it's actually a pattern-finding technique for mining inspiration from your data. Contrary to popular belief, it has nothing to do with machines running around without adult supervision, forming their own opinions about things.
An empirical study on evaluation metrics of generative adversarial networks
Xu, Qiantong, Huang, Gao, Yuan, Yang, Guo, Chuan, Sun, Yu, Wu, Felix, Weinberger, Kilian
Evaluating generative adversarial networks (GANs) is inherently challenging. In this paper, we revisit several representative sample-based evaluation metrics for GANs, and address the problem of how to evaluate the evaluation metrics. We start with a few necessary conditions for metrics to produce meaningful scores, such as distinguishing real from generated samples, identifying mode dropping and mode collapsing, and detecting overfitting. With a series of carefully designed experiments, we comprehensively investigate existing sample-based metrics and identify their strengths and limitations in practical settings. Based on these results, we observe that kernel Maximum Mean Discrepancy (MMD) and the 1-Nearest-Neighbor (1-NN) two-sample test seem to satisfy most of the desirable properties, provided that the distances between samples are computed in a suitable feature space. Our experiments also unveil interesting properties about the behavior of several popular GAN models, such as whether they are memorizing training samples, and how far they are from learning the target distribution.
Improving Consistency-Based Semi-Supervised Learning with Weight Averaging
Athiwaratkun, Ben, Finzi, Marc, Izmailov, Pavel, Wilson, Andrew Gordon
Recent advances in deep unsupervised learning have renewed interest in semi-supervised methods, which can learn from both labeled and unlabeled data. Presently the most successful approaches to semi-supervised learning are based on consistency regularization, whereby a model is trained to be robust to small perturbations of its inputs and parameters. We show that consistency regularization leads to flatter but narrower optima. We also show that the test error surface for these methods is approximately convex in regions of weight space traversed by SGD. Inspired by these observations, we propose to train consistency based semi-supervised models with stochastic weight averaging (SWA), a recent method which averages weights along the trajectory of SGD. We also develop fast-SWA, which further accelerates convergence by averaging multiple points within each cycle of a cyclical learning rate schedule. With fast-SWA we achieve the best known semi-supervised results on CIFAR-10 and CIFAR-100 over many different numbers of observed training labels. For example, we achieve 95.0% accuracy on CIFAR-10 with only 4000 labels, compared to the previous best result in the literature of 93.7%. We also improve the best known accuracy for domain adaptation from CIFAR-10 to STL from 80% to 83%. Finally, we show that with fast-SWA the simple $\Pi$ model becomes state-of-the-art for large labeled settings.
Writing the future of machine learning and invention
June's NAACL conference saw machine learning specialists from technology company Iprova present a paper introducing a new and effective method for the unsupervised training of machine learning algorithms to infer sentence embeddings. The NAACL (North American Chapter of the Association for Computational Linguistics) Human Language Technologies (HLT) conference took place at the Hyatt Regency New Orleans hotel, Louisiana, from June 1โ6, 2018. The research paper, entitled "Unsupervised Learning of Sentence Embeddings using Compositional n-Gram Features", will be presented by Matteo Pagliardini. Pagliardini is a senior machine learning engineer at Iprova and one of the three scientists that authored the research paper and developed the new model for unsupervised training, Sent2Vec. While there have been several successes in deep learning in recent years, the paper notes that these have almost exclusively relied on supervised training.
Semi-Supervised Learning via Compact Latent Space Clustering
Kamnitsas, Konstantinos, Castro, Daniel C., Folgoc, Loic Le, Walker, Ian, Tanno, Ryutaro, Rueckert, Daniel, Glocker, Ben, Criminisi, Antonio, Nori, Aditya
We present a novel cost function for semi-supervised learning of neural networks that encourages compact clustering of the latent space to facilitate separation. The key idea is to dynamically create a graph over embeddings of labeled and unlabeled samples of a training batch to capture underlying structure in feature space, and use label propagation to estimate its high and low density regions. We then devise a cost function based on Markov chains on the graph that regularizes the latent space to form a single compact cluster per class, while avoiding to disturb existing clusters during optimization. We evaluate our approach on three benchmarks and compare to state-of-the art with promising results. Our approach combines the benefits of graph-based regularization with efficient, inductive inference, does not require modifications to a network architecture, and can thus be easily applied to existing networks to enable an effective use of unlabeled data.
Learning Neural Random Fields with Inclusive Auxiliary Generators
In this paper we develop Neural Random Field learning with Inclusive-divergence minimized Auxiliary Generators (NRF-IAG), which is under-appreciated in the literature. The contributions are two-fold. First, we rigorously apply the stochastic approximation algorithm to solve the joint optimization and provide theoretical justification. The new approach of learning NRF-IAG achieves superior unsupervised learning performance competitive with state-of-the-art deep generative models (DGMs) in terms of sample generation quality. Second, semi-supervised learning (SSL) with NRF-IAG gives rise to strong classification results comparable to state-of-art DGM-based SSL methods, and simultaneously achieves superior generation. This is in contrast to the conflict of good classification and good generation, as observed in GAN-based SSL.
What Is Generative Adversarial Networks (GAN)?
The Generative Adversarial Networks (GAN) is, in fact, never a single network. It is a set of networks, at least two, operating at the same place but working against each other. Each of the networks brings its own unique set of results. For instance, in GAN approach, the first network creates realistic images, while the second one identifies whether those are real or not. It is like the first network is synthesizing something and the second one is monitoring its operations and controls what it creates.
Comprehensive Guide to Generative Adversarial Networks and Wasserstein GANs
The year 2017 was a period of scientific breakthroughs in deep learning, with the publication of numerous research papers. Every year seems like a big leap toward artificial general intelligence, or AGI. One exciting development involves generative modelling and the use of Wasserstein GANs (Generative Adversarial Networks). An influential paper on the topic has completely changed the approach to generative modelling, moving beyond the time when Ian Goodfellow published the original GAN paper. This paper differs from earlier work: the training algorithm is backed up by theory, and few examples exist where theory-justified papers gave good empirical results.
Unsupervised Learning of Artistic Styles with Archetypal Style Analysis
Wynen, Daan, Schmid, Cordelia, Mairal, Julien
In this paper, we introduce an unsupervised learning approach to automatically discover, summarize, and manipulate artistic styles from large collections of paintings. Our method is based on archetypal analysis, which is an unsupervised learning technique akin to sparse coding with a geometric interpretation. When applied to deep image representations from a collection of artworks, it learns a dictionary of archetypal styles, which can be easily visualized. After training the model, the style of a new image, which is characterized by local statistics of deep visual features, is approximated by a sparse convex combination of archetypes. This enables us to interpret which archetypal styles are present in the input image, and in which proportion. Finally, our approach allows us to manipulate the coefficients of the latent archetypal decomposition, and achieve various special effects such as style enhancement, transfer, and interpolation between multiple archetypes.