Staddon, Jessica
Distinguishing Scams and Fraud with Ensemble Learning
Chadalavada, Isha, Huang, Tianhui, Staddon, Jessica
Users increasingly query LLM-enabled web chatbots for help with scam defense. The Consumer Financial Protection Bureau's complaints database is a rich data source for evaluating LLM performance on user scam queries, but currently the corpus does not distinguish between scam and non-scam fraud. We developed an LLM ensemble approach to distinguishing scam and fraud CFPB complaints and describe initial findings regarding the strengths and weaknesses of LLMs in the scam defense context.
Assessment of LLM Responses to End-user Security Questions
Prakash, Vijay, Lee, Kevin, Bhattacharya, Arkaprabha, Huang, Danny Yuxing, Staddon, Jessica
Answering end user security questions is challenging. While large language models (LLMs) like GPT, LLAMA, and Gemini are far from error-free, they have shown promise in answering a variety of questions outside of security. We studied LLM performance in the area of end user security by qualitatively evaluating 3 popular LLMs on 900 systematically collected end user security questions. While LLMs demonstrate broad generalist ``knowledge'' of end user security information, there are patterns of errors and limitations across LLMs consisting of stale and inaccurate answers, and indirect or unresponsive communication styles, all of which impacts the quality of information received. Based on these patterns, we suggest directions for model improvement and recommend user strategies for interacting with LLMs when seeking assistance with security.
Can LLMs be Scammed? A Baseline Measurement Study
Sehwag, Udari Madhushani, Patel, Kelly, Mosca, Francesca, Ravi, Vineeth, Staddon, Jessica
Despite the importance of developing generative AI models that can effectively resist scams, current literature lacks a structured framework for evaluating their vulnerability to such threats. In this work, we address this gap by constructing a benchmark based on the FINRA taxonomy and systematically assessing Large Language Models' (LLMs') vulnerability to a variety of scam tactics. First, we incorporate 37 well-defined base scam scenarios reflecting the diverse scam categories identified by FINRA taxonomy, providing a focused evaluation of LLMs' scam detection capabilities. Second, we utilize representative proprietary (GPT-3.5, GPT-4) and open-source (Llama) models to analyze their performance in scam detection. Third, our research provides critical insights into which scam tactics are most effective against LLMs and how varying persona traits and persuasive techniques influence these vulnerabilities. We reveal distinct susceptibility patterns across different models and scenarios, underscoring the need for targeted enhancements in LLM design and deployment.
Around the Water Cooler: Shared Discussion Topics and Contact Closeness in Social Search
Komanduri, Saranga (Carnegie Mellon University) | Fang, Lujun (University of Michigan at Ann Arbor) | Huffaker, David (Google, Inc) | Staddon, Jessica (Google, Inc)
Search engines are now augmenting search results with social annotations, i.e., endorsements from users’ social network contacts. However, there is currently a dearth of published research on the effects of these annotations on user choice. This work investigates two research questions associated with annotations: 1) do some contacts affect user choice more than others, and 2) are annotations relevant across various information needs. We conduct a controlled experiment with 355 participants, using hypothetical searches and annotations, and elicit users’ choices. We find that domain contacts are preferred to close contacts, and this preference persists across a variety of information needs. Further, these contacts need not be experts and might be identified easily from conversation data.
The Web as a Privacy Lab
Chow, Richard (PARC) | Fang, Ji (PARC) | Golle, Philippe (PARC) | Staddon, Jessica (PARC)
The privacy dangers of data proliferation on the Web are well-known. Information on the Web has facilitated the deanonymization of anonymous bloggers, the de-sanitization of government records and the identification of individuals based on search engine queries. What has received less attention is Web-mining in support of privacy. In this position paper we argue that the very ability ofWeb data to breach privacy demonstrates its value as a laboratory for the detection of privacy breaches before they happen. In addition, we argue that privacy-invasive services may become privacy-respecting by mining publicly available Web data, with little decrease in performance and efficiency.