Goto

Collaborating Authors

 Moon, Lori


Multi-step Inference over Unstructured Data

arXiv.org Artificial Intelligence

The advent of Large Language Models (LLMs) and Generative AI has revolutionized natural language applications across various domains. However, high-stakes decision-making tasks in fields such as medical, legal and finance require a level of precision, comprehensiveness, and logical consistency that pure LLM or Retrieval-Augmented-Generation (RAG) approaches often fail to deliver. At Elemental Cognition (EC), we have developed a neuro-symbolic AI platform to tackle these problems. The platform integrates fine-tuned LLMs for knowledge extraction and alignment with a robust symbolic reasoning engine for logical inference, planning and interactive constraint solving. We describe Cora, a Collaborative Research Assistant built on this platform, that is designed to perform complex research and discovery tasks in high-stakes domains. This paper discusses the multi-step inference challenges inherent in such domains, critiques the limitations of existing LLM-based methods, and demonstrates how Cora's neuro-symbolic approach effectively addresses these issues. We provide an overview of the system architecture, key algorithms for knowledge extraction and formal reasoning, and present preliminary evaluation results that highlight Cora's superior performance compared to well-known LLM and RAG baselines.


GLUCOSE: GeneraLized and COntextualized Story Explanations

arXiv.org Artificial Intelligence

When humans read or listen, they make implicit commonsense inferences that frame their understanding of what happened and why. As a step toward AI systems that can build similar mental models, we introduce GLUCOSE, a large-scale dataset of implicit commonsense causal knowledge, encoded as causal mini-theories about the world, each grounded in a narrative context. To construct GLUCOSE, we drew on cognitive psychology to identify ten dimensions of causal explanation, focusing on events, states, motivations, and emotions. Each GLUCOSE entry includes a story-specific causal statement paired with an inference rule generalized from the statement. This paper details two concrete contributions: First, we present our platform for effectively crowdsourcing GLUCOSE data at scale, which uses semi-structured templates to elicit causal explanations. Using this platform, we collected 440K specific statements and general rules that capture implicit commonsense knowledge about everyday situations. Second, we show that existing knowledge resources and pretrained language models do not include or readily predict GLUCOSE's rich inferential content. However, when state-of-the-art neural models are trained on this knowledge, they can start to make commonsense inferences on unseen stories that match humans' mental models.


Modal Verbs in the Common Ground: Discriminating Among Actual and Nonactual Uses of Could and Would for Improved Text Interpretation

AAAI Conferences

Modal verbs occur in contexts which convey information about non-actual states of affairs as well as in contexts which convey information about the actual world of the discourse. Modeling the semantic interpretation of non-actual states of affairs is notoriously complicated, sometimes requiring modal logic, belief revision, non-monotonic reasoning, and multi-agent autoepistemic models. This work presents linguistic features which disambiguate those instances of the past tense modal verbs `could’ and `would’ which occur in contexts where the proposition in the scope of the modal is not true in the actual world of the discourse from those instances which presuppose or entail that an event in their scope occurred in the actual world of the discourse. It also illustrates the complexity of the role of modal verbs in semantic interpretation and, consequently, the limitations of state of the art inference systems with respect to modal verbs. The features suggested for improving modal verb interpretation are based on the analysis of corpus data and insights from the linguistic literature.