Goto

Collaborating Authors

 Davis, Ernest


Logical Formalizations of Commonsense Reasoning: A Survey

Journal of Artificial Intelligence Research

Commonsense reasoning is in principle a central problem in artificial intelligence, but it is a very difficult one. One approach that has been pursued since the earliest days of the field has been to encode commonsense knowledge as statements in a logic-based representation language and to implement commonsense reasoning as some form of logical inference. This paper surveys the use of logic-based representations of commonsense knowledge in artificial intelligence research.


Planning, Executing, and Evaluating the Winograd Schema Challenge

AI Magazine

The Winograd Schema Challenge was proposed by Hector Levesque in 2011 as an alternative to the Turing Test. Chief among its features is a simple question format that can span many commonsense knowledge domains. Questions are chosen so that they do not require specialized knoweldge or training, and are easy for humans to answer. This article details our plans to run the WSC and evaluate results.


How to Write Science Questions that Are Easy for People and Hard for Computers

AI Magazine

As a challenge problem for AI systems, I propose the use of hand-constructed multiple-choice tests, with problems that are easy for people but hard for computers. Specifically, I discuss techniques for constructing such problems at the level of a fourth-grade child and at the level of a high-school student. For the fourth grade level questions, I argue that questions that require the understanding of time, impossible or pointless scenarios, of causality, of the human body, or of sets of objects, and questions that require combining facts or require simple inductive arguments of indeterminate length can be chosen to be easy for people, and are likely to be hard for AI programs, in the current state of the art. For the high-school level, I argue that questions that relate the formal science to the realia of laboratory experiments or of real-world observations are likely to be easy for people and hard for AI programs.


Planning, Executing, and Evaluating the Winograd Schema Challenge

AI Magazine

Turing test turns out to be highly susceptible to systems that few people would wish to call intelligent. The Loebner Prize Competition (Christian 2011) is in particular associated with the development of chatterbots that are best viewed as successors to ELIZA (Weizenbaum 1966), the program that fooled people into thinking that they were talking to a human psychotherapist by cleverly turning a person's statements into questions of the sort a therapist would ask. The knowledge and inference that characterize conversations of substance -- for example, discussing alternate metaphors in sonnets of Shakespeare -- and which Turing presented as examples of the sorts of conversation that an intelligent system should be able to produce, are absent in these chatterbots. The focus is merely on engaging in surfacelevel conversation that can fool some humans who do not delve too deeply into a conversation, for at least a few minutes, into thinking that they are speaking to another person. The test taker, however, who is given a commonsense knowledge.


How to Write Science Questions that Are Easy for People and Hard for Computers

AI Magazine

As a challenge problem for AI systems, I propose the use of hand-constructed multiple-choice tests, with problems that are easy for people but hard for computers. Specifically, I discuss techniques for constructing such problems at the level of a fourth-grade child and at the level of a high-school student. For the fourth grade level questions, I argue that questions that require the understanding of time, impossible or pointless scenarios, of causality, of the human body, or of sets of objects, and questions that require combining facts or require simple inductive arguments of indeterminate length can be chosen to be easy for people, and are likely to be hard for AI programs, in the current state of the art. For the high-school level, I argue that questions that relate the formal science to the realia of laboratory experiments or of real-world observations are likely to be easy for people and hard for AI programs. I argue that these are more useful benchmarks than existing standardized tests such as the SATs or Regents tests. Since the questions in standardized tests are designed to be hard for people, they often leave many aspects of what is hard for computers but easy for people untested


The Limitations of Standardized Science Tests as Benchmarks for Artificial Intelligence Research: Position Paper

arXiv.org Artificial Intelligence

In this position paper, I argue that standardized tests for elementary science such as SAT or Regents tests are not very good benchmarks for measuring the progress of artificial intelligence systems in understanding basic science. The primary problem is that these tests are designed to test aspects of knowledge and ability that are challenging for people; the aspects that are challenging for AI systems are very different. In particular, standardized tests do not test knowledge that is obvious for people; none of this knowledge can be assumed in AI systems. Individual standardized tests also have specific features that are not necessarily appropriate for an AI benchmark. I analyze the Physics subject SAT in some detail and the New York State Regents Science test more briefly. I also argue that the apparent advantages offered by using standardized tests are mostly either minor or illusory. The one major real advantage is that the significance is easily explained to the public; but I argue that even this is a somewhat mixed blessing. I conclude by arguing that, first, more appropriate collections of exam style problems could be assembled, and second, that there are better kinds of benchmarks than exam-style problems. In an appendix I present a collection of sample exam-style problems that test kinds of knowledge missing from the standardized tests.


Reports of the AAAI 2011 Spring Symposia

AI Magazine

The Association for the Advancement of Artificial Intelligence, in cooperation with Stanford University's Department of Computer Science, presented the 2011 Spring Symposium Series Monday through Wednesday, March 21–23, 2011 at Stanford University. The titles of the eight symposia were AI and Health Communication, Artificial Intelligence and Sustainable Design, AI for Business Agility, Computational Physiology, Help Me Help You: Bridging the Gaps in Human-Agent Collaboration, Logical Formalizations of Commonsense Reasoning, Multirobot Systems and Physical Data Structures, and Modeling Complex Adaptive Systems As If They Were Voting Processes.


Reports of the AAAI 2011 Spring Symposia

AI Magazine

The titles of the eight symposia were Artificial Intelligence and Health Communication, Artificial Intelligence and Sustainable Design, Artificial Intelligence for Business Agility, Computational Physiology, Help Me Help You: Bridging the Gaps in Human-Agent Collaboration, Logical Formalizations of Commonsense Reasoning, Multirobot Systems and Physical Data Structures, and Modeling Complex Adaptive Systems As If They Were Voting Processes. The goal of the Artificial Intelligence and Health Communication symposium was to advance the conceptual design of automated systems that provide health services to patients and consumers through interdisciplinary insight from artificial intelligence, health communication and related areas of communication studies, discourse studies, public health, and psychology. There is a large and growing interest in the development of automated systems to provide health services to patients and consumers. In the last two decades, applications informed by research in health communication have been developed, for example, for promoting healthy behavior and for managing chronic diseases. While the value that these types of applications can offer to the community in terms of cost, access, and convenience is clear, there are still major challenges facing design of effective health communication systems. Overall, the participants found the format of the symposium engaging and constructive, and they The symposium was organized around five main expressed the desire to continue this initiative in concepts: (1) Patient empowerment and education further events.



Ontologies and Representations of Matter

AAAI Conferences

We carry out a comparative study of the expressive power of different ontologies of matter in terms of the ease with which simple physical knowledge can be represented. In particular, we consider five ontologies of models of matter: particle models, fields, two ontologies for continuous material, and a hybrid model. We evaluate these in terms of how easily eleven benchmark physical laws and scenarios can be represented.