Abductive Reasoning
Active Reasoning in an Open-World Environment
Xu, Manjie, Jiang, Guangyuan, Liang, Wei, Zhang, Chi, Zhu, Yixin
Recent advances in vision-language learning have achieved notable success on complete-information question-answering datasets through the integration of extensive world knowledge. Yet, most models operate passively, responding to questions based on pre-stored knowledge. In stark contrast, humans possess the ability to actively explore, accumulate, and reason using both newfound and existing information to tackle incomplete-information questions. In response to this gap, we introduce $Conan$, an interactive open-world environment devised for the assessment of active reasoning. $Conan$ facilitates active exploration and promotes multi-round abductive inference, reminiscent of rich, open-world settings like Minecraft. Diverging from previous works that lean primarily on single-round deduction via instruction following, $Conan$ compels agents to actively interact with their surroundings, amalgamating new evidence with prior knowledge to elucidate events from incomplete observations. Our analysis on $Conan$ underscores the shortcomings of contemporary state-of-the-art models in active exploration and understanding complex scenarios. Additionally, we explore Abduction from Deduction, where agents harness Bayesian rules to recast the challenge of abduction as a deductive process. Through $Conan$, we aim to galvanize advancements in active reasoning and set the stage for the next generation of artificial intelligence agents adept at dynamically engaging in environments.
- Asia > China > Beijing > Beijing (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Sweden > Skåne County > Malmö (0.04)
- Asia > Middle East > Jordan (0.04)
- Research Report > New Finding (0.46)
- Research Report > Promising Solution (0.34)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- (2 more...)
Neural Algorithmic Reasoning Without Intermediate Supervision
Rodionov, Gleb, Prokhorenkova, Liudmila
Neural algorithmic reasoning is an emerging area of machine learning focusing on building models that can imitate the execution of classic algorithms, such as sorting, shortest paths, etc. One of the main challenges is to learn algorithms that are able to generalize to out-of-distribution data, in particular with significantly larger input sizes. Recent work on this problem has demonstrated the advantages of learning algorithms step-by-step, giving models access to all intermediate steps of the original algorithm. In this work, we instead focus on learning neural algorithmic reasoning only from the input-output pairs without appealing to the intermediate supervision. We propose simple but effective architectural improvements and also build a self-supervised objective that can regularise intermediate computations of the model without access to the algorithm trajectory. We demonstrate that our approach is competitive to its trajectory-supervised counterpart on tasks from the CLRS Algorithmic Reasoning Benchmark and achieves new state-of-the-art results for several problems, including sorting, where we obtain significant improvements. Thus, learning without intermediate supervision is a promising direction for further research on neural reasoners.
- Information Technology > Artificial Intelligence > Machine Learning (0.87)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (0.40)
Axiomatic Aggregations of Abductive Explanations
Biradar, Gagan, Izza, Yacine, Lobo, Elita, Viswanathan, Vignesh, Zick, Yair
The recent criticisms of the robustness of post hoc model approximation explanation methods (like LIME and SHAP) have led to the rise of model-precise abductive explanations. For each data point, abductive explanations provide a minimal subset of features that are sufficient to generate the outcome. While theoretically sound and rigorous, abductive explanations suffer from a major issue -- there can be several valid abductive explanations for the same data point. In such cases, providing a single abductive explanation can be insufficient; on the other hand, providing all valid abductive explanations can be incomprehensible due to their size. In this work, we solve this issue by aggregating the many possible abductive explanations into feature importance scores. We propose three aggregation methods: two based on power indices from cooperative game theory and a third based on a well-known measure of causal strength. We characterize these three methods axiomatically, showing that each of them uniquely satisfies a set of desirable properties. We also evaluate them on multiple datasets and show that these explanations are robust to the attacks that fool SHAP and LIME.
- Asia > Singapore (0.04)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.04)
- North America > United States > Florida > Broward County (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.93)
DOMINO: A Dual-System for Multi-step Visual Language Reasoning
Wang, Peifang, Golovneva, Olga, Aghajanyan, Armen, Ren, Xiang, Chen, Muhao, Celikyilmaz, Asli, Fazel-Zarandi, Maryam
Visual language reasoning requires a system to extract text or numbers from information-dense images like charts or plots and perform logical or arithmetic reasoning to arrive at an answer. To tackle this task, existing work relies on either (1) an end-to-end vision-language model trained on a large amount of data, or (2) a two-stage pipeline where a captioning model converts the image into text that is further read by another large language model to deduce the answer. However, the former approach forces the model to answer a complex question with one single step, and the latter approach is prone to inaccurate or distracting information in the converted text that can confuse the language model. In this work, we propose a dual-system for multi-step multimodal reasoning, which consists of a "System-1" step for visual information extraction and a "System-2" step for deliberate reasoning. Given an input, System-2 breaks down the question into atomic sub-steps, each guiding System-1 to extract the information required for reasoning from the image. Experiments on chart and plot datasets show that our method with a pre-trained System-2 module performs competitively compared to prior work on in- and out-of-distribution data. By fine-tuning the System-2 module (LLaMA-2 70B) on only a small amount of data on multi-step reasoning, the accuracy of our method is further improved and surpasses the best fully-supervised end-to-end approach by 5.7% and a pipeline approach with FlanPaLM (540B) by 7.5% on a challenging dataset with human-authored questions.
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (0.53)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.53)
A Scientific Feud Breaks Out Into the Open
For years now, Hakwan Lau has suffered from an inner torment. Lau is a neuroscientist who studies the sense of awareness that all of us experience during our every waking moment. How this awareness arises from ordinary matter is an ancient mystery. Several scientific theories purport to explain it, and Lau feels that one of them, called integrated information theory (IIT), has received a disproportionate amount of media attention. He's annoyed that its proponents tout it as the dominant theory in the press.
- North America > United States (0.15)
- North America > Canada > Ontario > National Capital Region > Ottawa (0.05)
- Asia > Japan (0.05)
- Health & Medicine > Therapeutic Area > Neurology (0.51)
- Media > News (0.48)
- Information Technology > Communications > Social Media (0.70)
- Information Technology > Information Management (0.61)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (0.40)
Self-Consistent Narrative Prompts on Abductive Natural Language Inference
Chan, Chunkit, Liu, Xin, Chan, Tsz Ho, Cheng, Jiayang, Song, Yangqiu, Wong, Ginny, See, Simon
Abduction has long been seen as crucial for narrative comprehension and reasoning about everyday situations. The abductive natural language inference ($\alpha$NLI) task has been proposed, and this narrative text-based task aims to infer the most plausible hypothesis from the candidates given two observations. However, the inter-sentential coherence and the model consistency have not been well exploited in the previous works on this task. In this work, we propose a prompt tuning model $\alpha$-PACE, which takes self-consistency and inter-sentential coherence into consideration. Besides, we propose a general self-consistent framework that considers various narrative sequences (e.g., linear narrative and reverse chronology) for guiding the pre-trained language model in understanding the narrative context of input. We conduct extensive experiments and thorough ablation studies to illustrate the necessity and effectiveness of $\alpha$-PACE. The performance of our method shows significant improvement against extensive competitive baselines.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (7 more...)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.70)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.31)
Towards Controllable Natural Language Inference through Lexical Inference Types
Zhang, Yingji, Carvalho, Danilo S., Pratt-Hartmann, Ian, Freitas, Andre
Explainable natural language inference aims to provide a mechanism to produce explanatory (abductive) inference chains which ground claims to their supporting premises. A recent corpus called EntailmentBank strives to advance this task by explaining the answer to a question using an entailment tree \cite{dalvi2021explaining}. They employ the T5 model to directly generate the tree, which can explain how the answer is inferred. However, it lacks the ability to explain and control the generation of intermediate steps, which is crucial for the multi-hop inference process. % One recent corpus, EntailmentBank, aims to push this task forward by explaining an answer to a question according to an entailment tree \cite{dalvi2021explaining}. They employ T5 to generate the tree directly, which can explain how the answer is inferred but cannot explain how the intermediate is generated, which is essential to the multi-hop inference process. In this work, we focus on proposing a controlled natural language inference architecture for multi-premise explanatory inference. To improve control and enable explanatory analysis over the generation, we define lexical inference types based on Abstract Meaning Representation (AMR) graph and modify the architecture of T5 to learn a latent sentence representation (T5 bottleneck) conditioned on said type information. We also deliver a dataset of approximately 5000 annotated explanatory inference steps, with well-grounded lexical-symbolic operations. Experimental results indicate that the inference typing induced at the T5 bottleneck can help T5 to generate a conclusion under explicit control.
- North America > United States > New York > New York County > New York City (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > United Kingdom > England > Greater Manchester > Manchester (0.04)
- (3 more...)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.94)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.88)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (0.74)
Abductive Reasoning with the GPT-4 Language Model: Case studies from criminal investigation, medical practice, scientific research
This study evaluates the GPT-4 Large Language Model's abductive reasoning in complex fields like medical diagnostics, criminology, and cosmology. Using an interactive interview format, the AI assistant demonstrated reliability in generating and selecting hypotheses. It inferred plausible medical diagnoses based on patient data and provided potential causes and explanations in criminology and cosmology. The results highlight the potential of LLMs in complex problem-solving and the need for further research to maximize their practical applications. Keywords: GPT-4 Language Model, Abductive Reasoning, Medical Diagnostics, Criminology, Cosmology, Hypothesis Generation 1 Introduction The rise of Large Language Models (LLMs) like GPT-4 (OpenAI, 2023) has marked a significant milestone in artificial intelligence, demonstrating an exceptional ability to mimic human-like text. Yet, this progress has sparked intense discussions among scholars. The discourse is largely polarized between two perspectives: one, the critique that these models, often referred to as "stochastic parrots" (Bender et al., 2021), are devoid of true creativity, and two, the counter-argument that they possess an excessive degree of inventiveness often yielding outputs that veer more towards the realm of fantasy than fact. This article investigates these debates, specifically within the context of abductive reasoning, a field that demands a careful balance between creativity and constraint. Abductive reasoning, often called "inference to the best explanation," involves generating and evaluating hypotheses to explain observations.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.05)
- North America > United States > Alabama (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Italy > Molise > Campobasso Province > Campobasso (0.04)
- Research Report (1.00)
- Overview (1.00)
- Health & Medicine > Diagnostic Medicine (0.77)
- Law > Criminal Law (0.66)
- Health & Medicine > Therapeutic Area (0.47)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Diagrammatization: Rationalizing with diagrammatic AI explanations for abductive-deductive reasoning on hypotheses
Lim, Brian Y., Cahaly, Joseph P., Sng, Chester Y. F., Chew, Adam
Many visualizations have been developed for explainable AI (XAI), but they often require further reasoning by users to interpret. We argue that XAI should support diagrammatic and abductive reasoning for the AI to perform hypothesis generation and evaluation to reduce the interpretability gap. We propose Diagrammatization to i) perform Peircean abductive-deductive reasoning, ii) follow domain conventions, and iii) explain with diagrams visually or verbally. We implemented DiagramNet for a clinical application to predict cardiac diagnoses from heart auscultation, and explain with shape-based murmur diagrams. In modeling studies, we found that DiagramNet not only provides faithful murmur shape explanations, but also has better prediction performance than baseline models. We further demonstrate the interpretability and trustworthiness of diagrammatic explanations in a qualitative user study with medical students, showing that clinically-relevant, diagrammatic explanations are preferred over technical saliency map explanations. This work contributes insights into providing domain-conventional abductive explanations for user-centric XAI.
- Asia > Singapore > Central Region > Singapore (0.04)
- Asia > Middle East > Jordan (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (5 more...)
- Overview (0.92)
- Research Report > Experimental Study (0.67)
- Research Report > New Finding (0.45)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Therapeutic Area > Cardiology/Vascular Diseases (1.00)
- Health & Medicine > Diagnostic Medicine (1.00)
- Education (1.00)
- Information Technology > Data Science > Data Mining (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Abductive Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Explanation & Argumentation (1.00)
- (3 more...)
Do ghosts really exist? 5 possible scientific explanations for paranormal activity REVEALED
Whether we like to admit it or not, many of us have probably questioned if a bump in the night was actually a ghost at some point or another. And if you're really unlucky, you might even believe you've see a spirit in the flesh. But what exactly makes us feel like we are in the presence of something beyond the grave? Exploding head syndrome, sleep paralysis and even mould can be the source of a chill down your spine or the inkling that someone is watching. So, brace yourselves, as MailOnline explores five possible scientific explanations behind experiences of paranormal activity.
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.70)
- Health & Medicine > Therapeutic Area > Neurology (0.48)