Goto

Collaborating Authors

 Heckerman, David


The Lumiere Project: Bayesian User Modeling for Inferring the Goals and Needs of Software Users

arXiv.org Artificial Intelligence

The Lumiere Project centers on harnessing probability and utility to provide assistance to computer software users. We review work on Bayesian user models that can be employed to infer a users needs by considering a user's background, actions, and queries. Several problems were tackled in Lumiere research, including (1) the construction of Bayesian models for reasoning about the time-varying goals of computer users from their observed actions and queries, (2) gaining access to a stream of events from software applications, (3) developing a language for transforming system events into observational variables represented in Bayesian user models, (4) developing persistent profiles to capture changes in a user expertise, and (5) the development of an overall architecture for an intelligent user interface. Lumiere prototypes served as the basis for the Office Assistant in the Microsoft Office '97 suite of productivity applications.


Staged Mixture Modelling and Boosting

arXiv.org Machine Learning

In this paper, we introduce and evaluate a data-driven staged mixture modeling technique for building density, regression, and classification models. Our basic approach is to sequentially add components to a finite mixture model using the structural expectation maximization (SEM) algorithm. We show that our technique is qualitatively similar to boosting. This correspondence is a natural byproduct of the fact that we use the SEM algorithm to sequentially fit the mixture model. Finally, in our experimental evaluation, we demonstrate the effectiveness of our approach on a variety of prediction and density estimation tasks using real-world data.


Large-Sample Learning of Bayesian Networks is NP-Hard

arXiv.org Artificial Intelligence

In this paper, we provide new complexity results for algorithms that learn discrete-variable Bayesian networks from data. Our results apply whenever the learning algorithm uses a scoring criterion that favors the simplest model able to represent the generative distribution exactly. Our results therefore hold whenever the learning algorithm uses a consistent scoring criterion and is applied to a sufficiently large dataset. We show that identifying high-scoring structures is hard, even when we are given an independence oracle, an inference oracle, and/or an information oracle. Our negative results also apply to the learning of discrete-variable Bayesian networks in which each node has at most k parents, for all k > 3.


Using ``epitomes'' to model genetic diversity: Rational design of HIV vaccine cocktails

Neural Information Processing Systems

We introduce a new model of genetic diversity which summarizes a large input dataset into an epitome, a short sequence or a small set of short sequences of probability distributions capturing many overlapping subsequences from the dataset. The epitome as a representation has already been used in modeling real-valued signals, such as images and audio. The discrete sequence model we introduce in this paper targets applications in genetics, from multiple alignment to recombination and mutation inference. In our experiments, we concentrate on modeling the diversity of HIV where the epitome emerges as a natural model for producing relatively small vaccines covering a large number of immune system targets known as epitopes. Our experiments show that the epitome includes more epitopes than other vaccine designs of similar length, including cocktails of consensus strains, phylogenetic tree centers, and observed strains. We also discuss epitome designs that take into account uncertainty about T-cell cross reactivity and epitope presentation. In our experiments, we find that vaccine optimization is fairly robust to these uncertainties.


Using ``epitomes'' to model genetic diversity: Rational design of HIV vaccine cocktails

Neural Information Processing Systems

We introduce a new model of genetic diversity which summarizes a large input dataset into an epitome, a short sequence or a small set of short sequences of probability distributions capturing many overlapping subsequences fromthe dataset. The epitome as a representation has already been used in modeling real-valued signals, such as images and audio. The discrete sequence model we introduce in this paper targets applications in genetics, from multiple alignment to recombination and mutation inference. Inour experiments, we concentrate on modeling the diversity of HIV where the epitome emerges as a natural model for producing relatively smallvaccines covering a large number of immune system targets known as epitopes. Our experiments show that the epitome includes more epitopes than other vaccine designs of similar length, including cocktails of consensus strains, phylogenetic tree centers, and observed strains. We also discuss epitome designs that take into account uncertainty about T-cell cross reactivity and epitope presentation. In our experiments, we find that vaccine optimization is fairly robust to these uncertainties.


The First Conference on E-mail and Anti-Spam

AI Magazine

The First Conference on E-mail and Anti- Spam was held from July 30 to July 31, 2004 in Mountain View, California. The conference, attended by 180 researchers, featured 29 papers that covered a number of topics, including e-mail in general, nonstatistical techniques for stopping spam, machine learning techniques, issues of identity in e-mail, as well as law and policy. The 2005 conference will be held at Stanford University from July 21 to 22.


The First Conference on E-mail and Anti-Spam

AI Magazine

The First Conference on E-mail and Anti- Spam was held from July 30 to July 31, 2004 in Mountain View, California. The conference, attended by 180 researchers, featured 29 papers that covered a number of topics, including e-mail in general, nonstatistical techniques for stopping spam, machine learning techniques, issues of identity in e-mail, as well as law and policy. The 2005 conference will be held at Stanford University from July 21 to 22.


Thinking Backward for Knowledge Acquisition

AI Magazine

This article examines the direction in which knowledge bases are constructed for diagnosis and decision making. When building an expert system, it is traditional to elicit knowledge from an expert in the direction in which the knowledge is to be applied, namely, from observable evidence toward unobservable hypotheses. Therefore, we argue that a knowledge base be constructed following the expert's natural reasoning direction, and then reverse the direction for use. This choice of representation direction facilitates knowledge acquisition in deterministic domains and is essential when a problem involves uncertainty.


Thinking Backward for Knowledge Acquisition

AI Magazine

This article examines the direction in which knowledge bases are constructed for diagnosis and decision making. When building an expert system, it is traditional to elicit knowledge from an expert in the direction in which the knowledge is to be applied, namely, from observable evidence toward unobservable hypotheses. However, experts usually find it simpler to reason in the opposite direction-from hypotheses to unobservable evidence-because this direction reflects causal relationships. Therefore, we argue that a knowledge base be constructed following the expert's natural reasoning direction, and then reverse the direction for use. This choice of representation direction facilitates knowledge acquisition in deterministic domains and is essential when a problem involves uncertainty. We illustrate this concept with influence diagrams, a methodology for graphically representing a joint probability distribution. Influence diagrams provide a practical means by which an expert can characterize the qualitative and quantitative relationships among evidence and hypotheses in the apporiate direction. Once constructed, the relationships can easily be reserved into the less intuitive direction in order to perform inference inference and diagnosis. In this way, knowledge acquisition is made cognitively simple; the machine carries the burden of translating the representation.