Goto

Collaborating Authors

 amplify




Post Hoc Explanations of Language Models Can Improve Language Models

Neural Information Processing Systems

Large Language Models (LLMs) have demonstrated remarkable capabilities in performing complex tasks. Moreover, recent research has shown that incorporating human-annotated rationales (e.g., Chain-of-Thought prompting) during in-context learning can significantly enhance the performance of these models, particularly on tasks that require reasoning capabilities. However, incorporating such rationales poses challenges in terms of scalability as this requires a high degree of human involvement. In this work, we present a novel framework, Amplifying Model Performance by Leveraging In-Context Learning with Post Hoc Explanations (AMPLIFY), which addresses the aforementioned challenges by automating the process of rationale generation.



Post Hoc Explanations of Language Models Can Improve Language Models

Neural Information Processing Systems

Large Language Models (LLMs) have demonstrated remarkable capabilities in performing complex tasks. Moreover, recent research has shown that incorporating human-annotated rationales (e.g., Chain-of-Thought prompting) during in-context learning can significantly enhance the performance of these models, particularly on tasks that require reasoning capabilities. However, incorporating such rationales poses challenges in terms of scalability as this requires a high degree of human involvement. In this work, we present a novel framework, Amplifying Model Performance by Leveraging In-Context Learning with Post Hoc Explanations (AMPLIFY), which addresses the aforementioned challenges by automating the process of rationale generation. More specifically, we construct automated natural language rationales that embed insights from post hoc explanations to provide corrective signals to LLMs.


Towards Collective Superintelligence: Amplifying Group IQ using Conversational Swarms

Rosenberg, Louis, Willcox, Gregg, Schumann, Hans, Mani, Ganesh

arXiv.org Artificial Intelligence

Swarm Intelligence (SI) is a natural phenomenon that enables biological groups to amplify their combined intellect by forming real-time systems. Artificial Swarm Intelligence (or Swarm AI) is a technology that enables networked human groups to amplify their combined intelligence by forming similar systems. In the past, swarm-based methods were constrained to narrowly defined tasks like probabilistic forecasting and multiple-choice decision making. A new technology called Conversational Swarm Intelligence (CSI) was developed in 2023 that amplifies the decision-making accuracy of networked human groups through natural conversational deliberations. The current study evaluated the ability of real-time groups using a CSI platform to take a common IQ test known as Raven's Advanced Progressive Matrices (RAPM). First, a baseline group of participants took the Raven's IQ test by traditional survey. This group averaged 45.6% correct. Then, groups of approximately 35 individuals answered IQ test questions together using a CSI platform called Thinkscape. These groups averaged 80.5% correct. This places the CSI groups in the 97th percentile of IQ test-takers and corresponds to an effective IQ increase of 28 points (p<0.001). This is an encouraging result and suggests that CSI is a powerful method for enabling conversational collective intelligence in large, networked groups. In addition, because CSI is scalable across groups of potentially any size, this technology may provide a viable pathway to building a Collective Superintelligence.


AMPLIFY:Attention-based Mixup for Performance Improvement and Label Smoothing in Transformer

Yang, Leixin, Xiang, Yu

arXiv.org Artificial Intelligence

Mixup is an effective data augmentation method that generates new augmented samples by aggregating linear combinations of different original samples. However, if there are noises or aberrant features in the original samples, Mixup may propagate them to the augmented samples, leading to over-sensitivity of the model to these outliers . To solve this problem, this paper proposes a new Mixup method called AMPLIFY. This method uses the Attention mechanism of Transformer itself to reduce the influence of noises and aberrant values in the original samples on the prediction results, without increasing additional trainable parameters, and the computational cost is very low, thereby avoiding the problem of high resource consumption in common Mixup methods such as Sentence Mixup . The experimental results show that, under a smaller computational resource cost, AMPLIFY outperforms other Mixup methods in text classification tasks on 7 benchmark datasets, providing new ideas and new ways to further improve the performance of pre-trained models based on the Attention mechanism, such as BERT, ALBERT, RoBERTa, and GPT. Our code can be obtained at https://github.com/kiwi-lilo/AMPLIFY.


Post Hoc Explanations of Language Models Can Improve Language Models

Krishna, Satyapriya, Ma, Jiaqi, Slack, Dylan, Ghandeharioun, Asma, Singh, Sameer, Lakkaraju, Himabindu

arXiv.org Artificial Intelligence

Large Language Models (LLMs) have demonstrated remarkable capabilities in performing complex tasks. Moreover, recent research has shown that incorporating human-annotated rationales (e.g., Chain-of-Thought prompting) during in-context learning can significantly enhance the performance of these models, particularly on tasks that require reasoning capabilities. However, incorporating such rationales poses challenges in terms of scalability as this requires a high degree of human involvement. In this work, we present a novel framework, Amplifying Model Performance by Leveraging In-Context Learning with Post Hoc Explanations (AMPLIFY), which addresses the aforementioned challenges by automating the process of rationale generation. To this end, we leverage post hoc explanation methods which output attribution scores (explanations) capturing the influence of each of the input features on model predictions. More specifically, we construct automated natural language rationales that embed insights from post hoc explanations to provide corrective signals to LLMs. Extensive experimentation with real-world datasets demonstrates that our framework, AMPLIFY, leads to prediction accuracy improvements of about 10-25% over a wide range of tasks, including those where prior approaches which rely on human-annotated rationales such as Chain-of-Thought prompting fall short. Our work makes one of the first attempts at highlighting the potential of post hoc explanations as valuable tools for enhancing the effectiveness of LLMs. Furthermore, we conduct additional empirical analyses and ablation studies to demonstrate the impact of each of the components of AMPLIFY, which, in turn, leads to critical insights for refining in-context learning.


The Strange: Scifi Mars robots meet real-world bounded rationality

Robohub

Even with the addition of a strange mineral, robots still obey the principle of bounded rationality in artificial intelligence set forth by Herb Simon. I cover bounded rationality in my Science Robotics review (image courtesy of @SciRobotics) but I am adding some more details here. Did you like the Western True Grit? If yes to any or all of the above, The Strange by Nathan Ballingrud is for you! First off, let's talk about the book. The Strange is set in a counterfactual Confederate States of America colony on Mars circa 1930s, evocative of Ray Bradbury's The Martian Chronicles.


Computer says "No": The Case Against Empathetic Conversational AI

Curry, Alba, Curry, Amanda Cercas

arXiv.org Artificial Intelligence

It is important to note that Recent work in conversational AI has focused our argument applies to any use of empathetic AI on generating empathetic responses to users' (see also for example (Morris et al., 2018; De Carolis emotional states (e.g., Ide and Kawahara, 2022; et al., 2017)). What happens if the chatbot gets Svikhnushina et al., 2022; Zhu et al., 2022) as a way it right? There may be instances where a chatbot to increase or maintain engagement and rapport correctly identifies that a given situation is worthy with the user and to simulate intelligence. However, of praise and amplifies the pride of the user and these empathetic responses are problematic.