supportiveness
From Reddit to Generative AI: Evaluating Large Language Models for Anxiety Support Fine-tuned on Social Media Data
Kursuncu, Ugur, Padhi, Trilok, Sinha, Gaurav, Erol, Abdulkadir, Mandivarapu, Jaya Krishna, Larrison, Christopher R.
The critical shortage of mental health services due to workforce limitations and logistical barriers, especially in underserved areas designated by the Health Resources & Services Administration (HRSA) 1, highlights the urgent need for accessible and scalable solutions. Traditional services often fail to address the diverse needs of individuals experiencing anxiety, prompting many, especially younger populations, to seek alternative emotional and psychological support online. While digital platforms offer immediate access, unregulated online interactions, including those with generative AI, may disseminate misleading information or inappropriate advice, potentially exacerbating anxiety symptoms (Tobias & Ito, 2021). Despite the great potential of generative AI to supplement mental health services, its deployment poses potentially significant risks. Unlike clinical practitioners, LLMs are not inherently equipped to manage emotionally complex or vulnerable conversations, which are critical to therapeutic relationships that create positive clinical outcomes (Rogers, 1957; Wampold, 2015).
- North America > United States > Georgia > Clarke County > Athens (0.14)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- North America > United States > Illinois > Champaign County > Champaign (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.66)
Supportiveness-based Knowledge Rewriting for Retrieval-augmented Language Modeling
Qiao, Zile, Ye, Wei, Jiang, Yong, Mo, Tong, Xie, Pengjun, Li, Weiping, Huang, Fei, Zhang, Shikun
Retrieval-augmented language models (RALMs) have recently shown great potential in mitigating the limitations of implicit knowledge in LLMs, such as untimely updating of the latest expertise and unreliable retention of long-tail knowledge. However, since the external knowledge base, as well as the retriever, can not guarantee reliability, potentially leading to the knowledge retrieved not being helpful or even misleading for LLM generation. In this paper, we introduce Supportiveness-based Knowledge Rewriting (SKR), a robust and pluggable knowledge rewriter inherently optimized for LLM generation. Specifically, we introduce the novel concept of "supportiveness"--which represents how effectively a knowledge piece facilitates downstream tasks--by considering the perplexity impact of augmented knowledge on the response text of a white-box LLM. Based on knowledge supportiveness, we first design a training data curation strategy for our rewriter model, effectively identifying and filtering out poor or irrelevant rewrites (e.g., with low supportiveness scores) to improve data efficacy. We then introduce the direct preference optimization (DPO) algorithm to align the generated rewrites to optimal supportiveness, guiding the rewriter model to summarize augmented content that better improves the final response. Comprehensive evaluations across six popular knowledge-intensive tasks and four LLMs have demonstrated the effectiveness and superiority of SKR. With only 7B parameters, SKR has shown better knowledge rewriting capability over GPT-4, the current state-of-the-art general-purpose LLM.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- (6 more...)
FASTTRACK: Fast and Accurate Fact Tracing for LLMs
Chen, Si, Kang, Feiyang, Yu, Ning, Jia, Ruoxi
Fact tracing seeks to identify specific training examples that serve as the knowledge source for a given query. Existing approaches to fact tracing rely on assessing the similarity between each training sample and the query along a certain dimension, such as lexical similarity, gradient, or embedding space. However, these methods fall short of effectively distinguishing between samples that are merely relevant and those that actually provide supportive evidence for the information sought by the query. This limitation often results in suboptimal effectiveness. Moreover, these approaches necessitate the examination of the similarity of individual training points for each query, imposing significant computational demands and creating a substantial barrier for practical applications. This paper introduces FASTTRACK, a novel approach that harnesses the capabilities of Large Language Models (LLMs) to validate supportive evidence for queries and at the same time clusters the training database towards a reduced extent for LLMs to trace facts. Our experiments show that FASTTRACK substantially outperforms existing methods in both accuracy and efficiency, achieving more than 100\% improvement in F1 score over the state-of-the-art methods while being X33 faster than \texttt{TracIn}.
- Asia > Taiwan (0.14)
- North America > United States > Maryland (0.05)
- North America > United States > Virginia (0.04)
- (14 more...)
- Government > Regional Government (0.93)
- Health & Medicine (0.68)
Are Hard Examples also Harder to Explain? A Study with Human and Model-Generated Explanations
Saha, Swarnadeep, Hase, Peter, Rajani, Nazneen, Bansal, Mohit
Recent work on explainable NLP has shown that few-shot prompting can enable large pretrained language models (LLMs) to generate grammatical and factual natural language explanations for data labels. In this work, we study the connection between explainability and sample hardness by investigating the following research question - "Are LLMs and humans equally good at explaining data labels for both easy and hard samples?" We answer this question by first collecting human-written explanations in the form of generalizable commonsense rules on the task of Winograd Schema Challenge (Winogrande dataset). We compare these explanations with those generated by GPT-3 while varying the hardness of the test samples as well as the in-context samples. We observe that (1) GPT-3 explanations are as grammatical as human explanations regardless of the hardness of the test samples, (2) for easy examples, GPT-3 generates highly supportive explanations but human explanations are more generalizable, and (3) for hard examples, human explanations are significantly better than GPT-3 explanations both in terms of label-supportiveness and generalizability judgements. We also find that hardness of the in-context examples impacts the quality of GPT-3 explanations. Finally, we show that the supportiveness and generalizability aspects of human explanations are also impacted by sample hardness, although by a much smaller margin than models. Supporting code and data are available at https://github.com/swarnaHub/ExplanationHardness
- North America > United States > Washington > King County > Seattle (0.04)
- Europe > Italy > Tuscany > Florence (0.04)
Judgment Aggregation in Multi-Agent Argumentation
Awad, Edmond, Booth, Richard, Tohme, Fernando, Rahwan, Iyad
Given a set of conflicting arguments, there can exist multiple plausible opinions about which arguments should be accepted, rejected, or deemed undecided. We study the problem of how multiple such judgments can be aggregated. We define the problem by adapting various classical social-choice-theoretic properties for the argumentation domain. We show that while argument-wise plurality voting satisfies many properties, it fails to guarantee the collective rationality of the outcome, and struggles with ties. We then present more general results, proving multiple impossibility results on the existence of any good aggregation operator. After characterising the sufficient and necessary conditions for satisfying collective rationality, we study whether restricting the domain of argument-wise plurality voting to classical semantics allows us to escape the impossibility result. We close by listing graph-theoretic restrictions under which argument-wise plurality rule does produce collectively rational outcomes. In addition to identifying fundamental barriers to collective argument evaluation, our results open up the door for a new research agenda for the argumentation and computational social choice communities.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
- North America > United States (0.04)
- South America > Argentina (0.04)
- (4 more...)
Mean-Field Theory of Meta-Learning
We discuss here the mean-field theory for a cellular automata model of meta-learning. The meta-learning is the process of combining outcomes of individual learning procedures in order to determine the final decision with higher accuracy than any single learning method. Our method is constructed from an ensemble of interacting, learning agents, that acquire and process incoming information using various types, or different versions of machine learning algorithms. The abstract learning space, where all agents are located, is constructed here using a fully connected model that couples all agents with random strength values. The cellular automata network simulates the higher level integration of information acquired from the independent learning trials. The final classification of incoming input data is therefore defined as the stationary state of the meta-learning system using simple majority rule, yet the minority clusters that share opposite classification outcome can be observed in the system. Therefore, the probability of selecting proper class for a given input data, can be estimated even without the prior knowledge of its affiliation. The fuzzy logic can be easily introduced into the system, even if learning agents are build from simple binary classification machine learning algorithms by calculating the percentage of agreeing agents.
- Europe > Poland > Masovia Province > Warsaw (0.04)
- North America > United States (0.04)
- Europe > Spain > Catalonia > Barcelona Province > Barcelona (0.04)
- Europe > France (0.04)
- Health & Medicine > Therapeutic Area > Immunology (1.00)
- Health & Medicine > Pharmaceuticals & Biotechnology (0.94)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (0.93)
- Education (0.68)