Goto

Collaborating Authors

 linc



Robustness of Neurosymbolic Reasoners on First-Order Logic Problems

Bansal, Hannah, Kurniawan, Kemal, Frermann, Lea

arXiv.org Artificial Intelligence

Recent trends in NLP aim to improve reasoning capabilities in Large Language Models (LLMs), with key focus on generalization and robustness to variations in tasks. Counterfactual task variants introduce minimal but semantically meaningful changes to otherwise valid first-order logic (FOL) problem instances altering a single predicate or swapping roles of constants to probe whether a reasoning system can maintain logical consistency under perturbation. Previous studies showed that LLMs becomes brittle on counterfactual variations, suggesting that they often rely on spurious surface patterns to generate responses. In this work, we explore if a neurosymbolic (NS) approach that integrates an LLM and a symbolic logical solver could mitigate this problem. Experiments across LLMs of varying sizes show that NS methods are more robust but perform worse overall that purely neural methods. We then propose NSCoT that combines an NS method and Chain-of-Thought (CoT) prompting and demonstrate that while it improves performance, NSCoT still lags behind standard CoT. Our analysis opens research directions for future work.



In-silico biological discovery with large perturbation models

Miladinovic, Djordje, Höppe, Tobias, Chevalley, Mathieu, Georgiou, Andreas, Stuart, Lachlan, Mehrjou, Arash, Bantscheff, Marcus, Schölkopf, Bernhard, Schwab, Patrick

arXiv.org Artificial Intelligence

Data generated in perturbation experiments link perturbations to the changes they elicit and therefore contain information relevant to numerous biological discovery tasks -- from understanding the relationships between biological entities to developing therapeutics. However, these data encompass diverse perturbations and readouts, and the complex dependence of experimental outcomes on their biological context makes it challenging to integrate insights across experiments. Here, we present the Large Perturbation Model (LPM), a deep-learning model that integrates multiple, heterogeneous perturbation experiments by representing perturbation, readout, and context as disentangled dimensions. LPM outperforms existing methods across multiple biological discovery tasks, including in predicting post-perturbation transcriptomes of unseen experiments, identifying shared molecular mechanisms of action between chemical and genetic perturbations, and facilitating the inference of gene-gene interaction networks.


Enhancing In-context Learning via Linear Probe Calibration

Abbas, Momin, Zhou, Yi, Ram, Parikshit, Baracaldo, Nathalie, Samulowitz, Horst, Salonidis, Theodoros, Chen, Tianyi

arXiv.org Artificial Intelligence

In-context learning (ICL) is a new paradigm for natural language processing that utilizes Generative Pre-trained Transformer (GPT)-like models. This approach uses prompts that include in-context demonstrations to generate the corresponding output for a new query input. However, applying ICL in real cases does not scale with the number of samples, and lacks robustness to different prompt templates and demonstration permutations. In this paper, we first show that GPT-like models using ICL result in unreliable predictions based on a new metric based on Shannon entropy. Then, to solve this problem, we propose a new technique called the Linear Probe Calibration (LinC), a method that calibrates the model's output probabilities, resulting in reliable predictions and improved performance, while requiring only minimal additional samples (as few as five labeled data samples). LinC significantly enhances the ICL test performance of GPT models on various benchmark datasets, with an average improvement of up to 21%, and up to a 50% improvement in some cases, and significantly boosts the performance of PEFT methods, especially in the low resource regime. Moreover, LinC achieves lower expected calibration error, and is highly robust to varying label proportions, prompt templates, and demonstration permutations. Our code is available at \url{https://github.com/mominabbass/LinC}.


LINC: A Neurosymbolic Approach for Logical Reasoning by Combining Language Models with First-Order Logic Provers

Olausson, Theo X., Gu, Alex, Lipkin, Benjamin, Zhang, Cedegao E., Solar-Lezama, Armando, Tenenbaum, Joshua B., Levy, Roger

arXiv.org Artificial Intelligence

Logical reasoning, i.e., deductively inferring the truth value of a conclusion from a set of premises, is an important task for artificial intelligence with wide potential impacts on science, mathematics, and society. While many prompting-based strategies have been proposed to enable Large Language Models (LLMs) to do such reasoning more effectively, they still appear unsatisfactory, often failing in subtle and unpredictable ways. In this work, we investigate the validity of instead reformulating such tasks as modular neurosymbolic programming, which we call LINC: Logical Inference via Neurosymbolic Computation. In LINC, the LLM acts as a semantic parser, translating premises and conclusions from natural language to expressions in first-order logic. These expressions are then offloaded to an external theorem prover, which symbolically performs deductive inference. Leveraging this approach, we observe significant performance gains on FOLIO and a balanced subset of ProofWriter for three different models in nearly all experimental conditions we evaluate. On ProofWriter, augmenting the comparatively small open-source StarCoder+ (15.5B parameters) with LINC even outperforms GPT-3.5 and GPT-4 with Chain-of-Thought (CoT) prompting by an absolute 38% and 10%, respectively. When used with GPT-4, LINC scores 26% higher than CoT on ProofWriter while performing comparatively on FOLIO. Further analysis reveals that although both methods on average succeed roughly equally often on this dataset, they exhibit distinct and complementary failure modes. We thus provide promising evidence for how logical reasoning over natural language can be tackled through jointly leveraging LLMs alongside symbolic provers. All corresponding code is publicly available at https://github.com/benlipkin/linc


Machine Learning & AI in the Classroom

#artificialintelligence

It is clear that the pandemic has had a dramatic impact on education, which for many has meant an unplanned and rapid move to online and blended learning approaches. In the Summer of 2021 a'Machine Learning & AI' module was developed by PhD student Joyce Mahon at UCD in a collaboration between the SFI CRT in Machine Learning and industry partner Huawei. Joyce is supervised by Dr. Brett Becker and Dr. Brian Mac Namee of the UCD School of Computer Science; and for the duration of this project worked alongside Dr. Keith Quille of TU Dublin, and with student volunteers. 'Machine Learning & AI' module was added to the CS_LINC platform developed in 2020 by the CS_INC team in TU Dublin. CS_LINC provides formal computer science curricula through free and easily accessible online modules.


How To Avoid A #ChatbotFail

#artificialintelligence

While brands have scrambled to launch Facebook Messenger chatbots since the social media behemoth opened up the channel for development last year, the early results haven't been particularly promising. Facebook is seeing a 70% failure rate among those 35,000 or so bots when it comes to understanding user requests. To combat this poor performance, Facebook is making some changes to Messenger, including adding a persistent menu that will allow users to choose from a number of requests or statements instead of using natural language and risking stumping the bot entirely. There's no question that AI will play a huge role in the future of retail, but in these early days of chatbots and virtual assistants, how do you reap the benefits while avoiding the pitfalls of this emerging technology? We caught up with Linc engineer Alessandro Sanchez to talk about the potential weaknesses in current chatbots and how smart brands are creating a chatbot experience that beats the odds and delivers great service.