Davis, Jesse
Conditioning in First-Order Knowledge Compilation and Lifted Probabilistic Inference
Broeck, Guy Van den (KU Leuven) | Davis, Jesse (KU Leuven)
Knowledge compilation is a powerful technique for compactly representing and efficiently reasoning about logical knowledge bases. It has been successfully applied to numerous problems in artificial intelligence, such as probabilistic inference and conformant planning. Conditioning, which updates a knowledge base with observed truth values for some propositions, is one of the fundamental operations employed for reasoning. In the propositional setting, conditioning can be efficiently applied in all cases. Recently, people have explored compilation for first-order knowledge bases. The majority of this work has centered around using first-order d-DNNF circuits as the target compilation language. However, conditioning has not been studied in this setting. This paper explores how to condition a first-order d-DNNF circuit. We show that it is possible to efficiently condition these circuits on unary relations. However, we prove that conditioning on higher arity relations is #P-hard. We study the implications of these findings on the application of performing lifted inference for first-order probabilistic models.This leads to a better understanding of which types of queries lifted inference can address.
Unachievable Region in Precision-Recall Space and Its Effect on Empirical Evaluation
Boyd, Kendrick, Costa, Vitor Santos, Davis, Jesse, Page, David
Precision-recall (PR) curves and the areas under them are widely used to summarize machine learning results, especially for data sets exhibiting class skew. They are often used analogously to ROC curves and the area under ROC curves. It is known that PR curves vary as class skew changes. What was not recognized before this paper is that there is a region of PR space that is completely unachievable, and the size of this region depends only on the skew. This paper precisely characterizes the size of that region and discusses its implications for empirical evaluation methodology in machine learning.
Demand-Driven Clustering in Relational Domains for Predicting Adverse Drug Events
Davis, Jesse, Costa, Vitor Santos, Peissig, Peggy, Caldwell, Michael, Berg, Elizabeth, Page, David
Learning from electronic medical records (EMR) is challenging due to their relational nature and the uncertain dependence between a patient's past and future health status. Statistical relational learning is a natural fit for analyzing EMRs but is less adept at handling their inherent latent structure, such as connections between related medications or diseases. One way to capture the latent structure is via a relational clustering of objects. We propose a novel approach that, instead of pre-clustering the objects, performs a demand-driven clustering during learning. We evaluate our algorithm on three real-world tasks where the goal is to use EMRs to predict whether a patient will have an adverse reaction to a medication. We find that our approach is more accurate than performing no clustering, pre-clustering, and using expert-constructed medical heterarchies.
Deep Transfer: A Markov Logic Approach
Davis, Jesse (Katholieke Universiteit Leuven) | Domingos, Pedro (University of Washington)
This article argues that currently the largest gap between human and machine learning is learning algorithms' inability to perform deep transfer, that is, generalize from one domain to another domain containing different objects, classes, properties and relations. We argue that second-order Markov logic is ideally suited for this purpose, and propose an approach based on it. Our algorithm discovers structural regularities in the source domain in the form of Markov logic formulas with predicate variables, and instantiates these formulas with predicates from the target domain. Our approach has successfully transferred learned knowledge among molecular biology, Web and social network domains.
Deep Transfer: A Markov Logic Approach
Davis, Jesse (Katholieke Universiteit Leuven) | Domingos, Pedro (University of Washington)
This article argues that currently the largest gap between human and machine learning is learning algorithms' inability to perform deep transfer, that is, generalize from one domain to another domain containing different objects, classes, properties and relations. We argue that second-order Markov logic is ideally suited for this purpose, and propose an approach based on it. Our algorithm discovers structural regularities in the source domain in the form of Markov logic formulas with predicate variables, and instantiates these formulas with predicates from the target domain. Our approach has successfully transferred learned knowledge among molecular biology, Web and social network domains.