Naïve Bayesian text classifiers are fast, accurate, simple, and easy to implement. In this article, I present a complete naïve Bayesian text classifier written in 100 lines of commented, nonobfuscated Perl.
This book (originally published in 1994 by Ellis Horwood) is now out of print. The copyright now resides with the editors who have decided to make the material freely available on the web.
This book is based on the EC (ESPRIT) project StatLog which compare and evaluated a range of classification techniques, with an assessment of their merits, disadvantages and range of application. This integrated volume provides a concise introduction to each method, and reviews comparative trials in large-scale commercial and industrial problems. It makes accessible to a wide range of workers the complex issue of classification as approached through machine learning, statistics and neural networks, encouraging a cross-fertilization between these discplines.
The year 2013 marks the 250th anniversary of Bayes rule, one of the two fundamental inferential principles of mathematical statistics. The rule has been influential over the entire period, and controversial over most of it. Its reliance on prior beliefs has been challenged by frequentism, which focuses instead on the behavior of specific estimates and tests under repeated use. Twentieth-century statistics was overwhelmingly behavioristic, especially in applications, but the twenty-first century has seen a resurgence of Bayesian ism. Some simple examples are used to show what’s at stake in the argument. The bootstrap, a computer-intensive inference machine, helps connect Bayesian and frequentist practice, leading finally to an empirical Bayes example of collaboration between the two philosophies.
NewsFinder automates the steps involved in finding, selecting, categorizing, and publishing news stories that meet relevance criteria for the Artificial Intelligence community. The software combines a broad search of online news sources with topic-specific trained models and heuristics. Since August 2010, the program has been used to operate the AI in the News service that is part of the AAAI AITopics website.
During the past decade has been an explosion in computation and information technology. With it has come vast amounts of data in a variety of fields such as medicine, biology, finance, and marketing. The challenge of understanding these data has led to the development of new tools in the field of statistics, and spawned new areas such as data mining, machine learning, and bioinformatics. Many of these tools have common underpinnings but are often expressed with different terminology. This book descibes the important ideas in these areas in a common conceptual framework. While the approach is statistical, the emphasis is on concepts rather than mathematics.
Each of the authors is an expert in machine learning / prediction, and in some cases invented the techniques we turn to today to make sense of big data: ensemble learning methods, penalized regression, additive models and nonparemetric smoothing, and much much more.
Full text available for a fee.
This book is a practical guide to classification learning systems and their applications. These computer programs learn from sample data and make predictions for new cases, sometimes exceeding the performance of humans.
Practical learning systems from statistical pattern recognition, neural networks, and machine learning are presented. The authors examine prominent methods from each area, using an engineering approach and taking the practitioner's viewpoint. Intuitive explanations with a minimum of mathematics make the material accessible to anyone--regardless of experience or special interests.
The underlying concepts of the learning methods are discussed with fully worked-out examples: their strengths and weaknesses, and the estimation of their future performance on specific applications. Throughout, the authors offer their own recommendations for selecting and applying learning methods such as linear discriminants, back-propagation neural networks, or decision trees. Learning systems are then contrasted with their rule-based counterparts from expert systems.
"The Machine Learning and Applied Statistics (MLAS) group is focused on learning from data and data mining. By building software that automatically learns from data, we enable applications that (1) do intelligent tasks such as handwriting recognition and natural-language processing, and (2) help human data analysts more easily explore and better understand their data."
The following links point to a set of tutorials on many aspects of statistical data mining, including the foundations of probability, the foundations of statistical data analysis, and most of the classic machine learning and data mining algorithms.These include classification algorithms such as decision trees, neural nets, Bayesian classifiers, Support Vector Machines and cased-based (aka non-parametric) learning. They include regression algorithms such as multivariate polynomial regression, MARS, Locally Weighted Regression, GMDH and neural nets. And they include other data mining operations such as clustering (mixture models, k-means and hierarchical), Bayesian networks and Reinforcement Learning.
The following links point to a set of tutorials on many aspects of statistical data mining, including the foundations of probability, the foundations of statistical data analysis, and most of the classic machine learning and data mining algorithms.
These include classification algorithms such as decision trees, neural nets, Bayesian classifiers, Support Vector Machines and cased-based (aka non-parametric) learning. They include regression algorithms such as multivariate polynomial regression, MARS, Locally Weighted Regression, GMDH and neural nets. And they include other data mining operations such as clustering (mixture models, k-means and hierarchical), Bayesian networks and Reinforcement Learning.
The active learning algorithm is faster and more accurate in guessing the age of an individual than conventional algorithms. They have, for example, developed computer algorithms for facial age classification -- the automated assignment of individuals to predefined age groups based on their facial features as seen on video captures or still images. A person can teach a computer to make better guesses by running its algorithm through a large database of facial images of which the age is known using sets of labeled images, but acquiring such a database can be both time-consuming and expensive. The technology could find use, for example, in digital signage where the machine determines the age group of the viewer and displays targeted advertisements designed for those age groups, or in interactive games where the machine automatically presents different games based on the players' age range.