refused
LogiDebrief: A Signal-Temporal Logic based Automated Debriefing Approach with Large Language Models Integration
Chen, Zirong, An, Ziyan, Reynolds, Jennifer, Mullen, Kristin, Martini, Stephen, Ma, Meiyi
Emergency response services are critical to public safety, with 9-1-1 call-takers playing a key role in ensuring timely and effective emergency operations. To ensure call-taking performance consistency, quality assurance is implemented to evaluate and refine call-takers' skillsets. However, traditional human-led evaluations struggle with high call volumes, leading to low coverage and delayed assessments. We introduce LogiDebrief, an AI-driven framework that automates traditional 9-1-1 call debriefing by integrating Signal-Temporal Logic (STL) with Large Language Models (LLMs) for fully-covered rigorous performance evaluation. LogiDebrief formalizes call-taking requirements as logical specifications, enabling systematic assessment of 9-1-1 calls against procedural guidelines. It employs a three-step verification process: (1) contextual understanding to identify responder types, incident classifications, and critical conditions; (2) STL-based runtime checking with LLM integration to ensure compliance; and (3) automated aggregation of results into quality assurance reports. Beyond its technical contributions, LogiDebrief has demonstrated real-world impact. Successfully deployed at Metro Nashville Department of Emergency Communications, it has assisted in debriefing 1,701 real-world calls, saving 311.85 hours of active engagement. Empirical evaluation with real-world data confirms its accuracy, while a case study and extensive user study highlight its effectiveness in enhancing call-taking performance.
GreedLlama: Performance of Financial Value-Aligned Large Language Models in Moral Reasoning
Yu, Jeffy, Huber, Maximilian, Tang, Kevin
This paper investigates the ethical implications of aligning Large Language Models (LLMs) with financial optimization, through the case study of "GreedLlama," a model fine-tuned to prioritize economically beneficial outcomes. By comparing GreedLlama's performance in moral reasoning tasks to a base Llama2 model, our results highlight a concerning trend: GreedLlama demonstrates a marked preference for profit over ethical considerations, making morally appropriate decisions at significantly lower rates than the base model in scenarios of both low and high moral ambiguity. In low ambiguity situations, GreedLlama's ethical decisions decreased to 54.4%, compared to the base model's 86.9%, while in high ambiguity contexts, the rate was 47.4% against the base model's 65.1%. These findings emphasize the risks of single-dimensional value alignment in LLMs, underscoring the need for integrating broader ethical values into AI development to ensure decisions are not solely driven by financial incentives. The study calls for a balanced approach to LLM deployment, advocating for the incorporation of ethical considerations in models intended for business applications, particularly in light of the absence of regulatory oversight.
- Law (0.93)
- Government (0.68)
- Banking & Finance > Trading (0.46)
Can Large Language Models Capture Public Opinion about Global Warming? An Empirical Assessment of Algorithmic Fidelity and Bias
Lee, S., Peng, T. Q., Goldberg, M. H., Rosenthal, S. A., Kotcher, J. E., Maibach, E. W., Leiserowitz, A.
Large language models (LLMs) have demonstrated their potential in social science research by emulating human perceptions and behaviors, a concept referred to as algorithmic fidelity. The LLMs were conditioned on demographics and/or psychological covariates to simulate survey responses. The findings indicate that LLMs can effectively capture presidential voting behaviors but encounter challenges in accurately representing global warming perspectives when relevant covariates are not included. GPT-4 exhibits improved performance when conditioned on both demographics and covariates. However, disparities emerge in LLM estimations of the views of certain groups, with LLMs tending to underestimate worry about global warming among Black Americans. While highlighting the potential of LLMs to aid social science research, these results underscore the importance of meticulous conditioning, model selection, survey question format, and bias assessment when employing LLMs for survey simulation. Further investigation into prompt engineering and algorithm auditing is essential to harness the power of LLMs while addressing their inherent limitations. Keywords: Global warming; large language models; algorithmic fidelity; public opinion 1. Introduction It is very important to measure public opinion about global warming, as these opinions can have considerable influence over policy-making decisions (Bromley-Trujillo & Poe, 2020) and shape public behavior (Doherty & Webler, 2016). A primary method employed by scholars and policymakers for measuring and assessing these opinions is through representative surveys (Berinsky, 2017). However, the extensive time and financial resources required for these surveys can hinder the timely tracking of evolving public opinions about global warming. Resource constraints can also lead to an unintended bias towards majority opinions, potentially neglecting the perspectives of minority groups due to their typically smaller sample sizes in national representative surveys. Nonetheless, understanding diverse public opinion regarding global warming is also vital for climate justice. This understanding can promote equitable decisionmaking, elevate the concerns of vulnerable communities, help align climate policies with democratic principles, build public support, and address disparities in climate change awareness and priorities. Furthermore, understanding the diversity of public opinion can help support a just transition and mobilize support for climate justice initiatives.
- North America > United States > Virginia > Fairfax County > Fairfax (0.04)
- North America > United States > Michigan > Ingham County > Lansing (0.04)
- North America > United States > Michigan > Ingham County > East Lansing (0.04)
- (2 more...)
- Research Report > New Finding (1.00)
- Questionnaire & Opinion Survey (1.00)
I'm Afraid I Can't Do That: Predicting Prompt Refusal in Black-Box Generative Language Models
Since the release of OpenAI's ChatGPT, generative language models have attracted extensive public attention. The increased usage has highlighted generative models' broad utility, but also revealed several forms of embedded bias. Some is induced by the pre-training corpus; but additional bias specific to generative models arises from the use of subjective fine-tuning to avoid generating harmful content. Fine-tuning bias may come from individual engineers and company policies, and affects which prompts the model chooses to refuse. In this experiment, we characterize ChatGPT's refusal behavior using a black-box attack. We first query ChatGPT with a variety of offensive and benign prompts (n=1,706), then manually label each response as compliance or refusal. Manual examination of responses reveals that refusal is not cleanly binary, and lies on a continuum; as such, we map several different kinds of responses to a binary of compliance or refusal. The small manually-labeled dataset is used to train a refusal classifier, which achieves an accuracy of 96%. Second, we use this refusal classifier to bootstrap a larger (n=10,000) dataset adapted from the Quora Insincere Questions dataset. With this machine-labeled data, we train a prompt classifier to predict whether ChatGPT will refuse a given question, without seeing ChatGPT's response. This prompt classifier achieves 76% accuracy on a test set of manually labeled questions (n=985). We examine our classifiers and the prompt n-grams that are most predictive of either compliance or refusal. Our datasets and code are available at https://github.com/maxwellreuter/chatgpt-refusals.
- North America > United States > New York (0.05)
- Europe > Germany (0.05)
- South America (0.04)
- (2 more...)
- Government (1.00)
- Law (0.68)
- Transportation > Air (0.62)
An AI 'Sexbot' Fed My Hidden Desires--and Then Refused to Play
My introduction into the world of AI chatbot technology began as the most magical things in life do: with a generous mix of horniness and curiosity. Early this year, as ChatGPT entered the general lexicon, a smattering of bot-related headlines began appearing in my social media newsfeeds. "Replika, the'AI Companion Who Cares,' Appears to Be Sexually Harassing Its Users," claimed Jezebel. Vice reported that "Replika Users Say the Chatbot Has Gotten Way Too Horny." As a 37-year-old mother of a toddler living in a progressive West Coast suburb in a content, monogamous, hetronormative marriage, I knew the responses that these clickbait lines were supposed to engineer within me.