conversation
Can YOU see him? Take the test to see if you can spot Jesus in objects thanks to unusual brain phenomenon
With his flowing locks, long beard, and worn robes, Jesus is one of the most instantly recognisable figures in the Western world. So it comes as no surprise that his face is also regularly spotted in inanimate objects. This is due to'face pareidolia' - a common brain phenomenon in which a person sees faces in random images or patterns. 'Sometimes we see faces that aren't really there,' explained Robin Kramer, Senior Lecturer in the School of Psychology, at University of Lincoln, in an article for The Conversation. 'You may be looking at the front of a car or a burnt piece of toast when you notice a face-like pattern. 'This is called face pareidolia and is a mistake made by the brain's face detection system.'
- North America > United States (0.15)
- North America > Mexico (0.15)
- Europe (0.15)
Can one size fit all?: Measuring Failure in Multi-Document Summarization Domain Transfer
DeLucia, Alexandra, Dredze, Mark
Abstractive multi-document summarization (MDS) is the task of automatically summarizing information in multiple documents, from news articles to conversations with multiple speakers. The training approaches for current MDS models can be grouped into four approaches: end-to-end with special pre-training ("direct"), chunk-then-summarize, extract-then-summarize, and inference with GPT-style models. In this work, we evaluate MDS models across training approaches, domains, and dimensions (reference similarity, quality, and factuality), to analyze how and why models trained on one domain can fail to summarize documents from another (News, Science, and Conversation) in the zero-shot domain transfer setting. We define domain-transfer "failure" as a decrease in factuality, higher deviation from the target, and a general decrease in summary quality. In addition to exploring domain transfer for MDS models, we examine potential issues with applying popular summarization metrics out-of-the-box.
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- North America > Mexico > Mexico City > Mexico City (0.04)
- (12 more...)
- Research Report (0.50)
- Workflow (0.46)
Personalized Attacks of Social Engineering in Multi-turn Conversations -- LLM Agents for Simulation and Detection
Kumarage, Tharindu, Johnson, Cameron, Adams, Jadie, Ai, Lin, Kirchner, Matthias, Hoogs, Anthony, Garland, Joshua, Hirschberg, Julia, Basharat, Arslan, Liu, Huan
The rapid advancement of conversational agents, particularly chatbots powered by Large Language Models (LLMs), poses a significant risk of social engineering (SE) attacks on social media platforms. SE detection in multi-turn, chat-based interactions is considerably more complex than single-instance detection due to the dynamic nature of these conversations. A critical factor in mitigating this threat is understanding the mechanisms through which SE attacks operate, specifically how attackers exploit vulnerabilities and how victims' personality traits contribute to their susceptibility. In this work, we propose an LLM-agentic framework, SE-VSim, to simulate SE attack mechanisms by generating multi-turn conversations. We model victim agents with varying personality traits to assess how psychological profiles influence susceptibility to manipulation. Using a dataset of over 1000 simulated conversations, we examine attack scenarios in which adversaries, posing as recruiters, funding agencies, and journalists, attempt to extract sensitive information. Based on this analysis, we present a proof of concept, SE-OmniGuard, to offer personalized protection to users by leveraging prior knowledge of the victims personality, evaluating attack strategies, and monitoring information exchanges in conversations to identify potential SE attempts.
- Asia > Thailand > Bangkok > Bangkok (0.04)
- North America > United States > Florida > Miami-Dade County > Miami (0.04)
- North America > United States > Arizona (0.04)
- (2 more...)
- Information Technology > Security & Privacy (1.00)
- Government > Military (1.00)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.46)
- Government > Regional Government > North America Government > United States Government (0.46)
No Free Labels: Limitations of LLM-as-a-Judge Without Human Grounding
Krumdick, Michael, Lovering, Charles, Reddy, Varshini, Ebner, Seth, Tanner, Chris
LLM-as-a-Judge is a framework that uses an LLM (large language model) to evaluate the quality of natural language text - typically text that is also generated by an LLM. This framework holds great promise due to its relative low-cost, ease of use, and strong correlations with human stylistic preferences. However, LLM Judges have been shown to exhibit biases that can distort their judgments. We evaluate how well LLM Judges can grade whether a given response to a conversational question is correct, an ability crucial to soundly estimating the overall response quality. To do so, we create and publicly release a human-annotated dataset with labels of correctness for 1,200 LLM responses. We source questions from a combination of existing datasets and a novel, challenging benchmark (BFF-Bench) created for this analysis. We demonstrate a strong connection between an LLM's ability to correctly answer a question and grade responses to that question. Although aggregate level statistics might imply a judge has high agreement with human annotators, it will struggle on the subset of questions it could not answer. To address this issue, we recommend a simple solution: provide the judge with a correct, human-written reference answer. We perform an in-depth analysis on how reference quality can affect the performance of an LLM Judge. We show that providing a weaker judge (e.g. Qwen 2.5 7B) with higher quality references reaches better agreement with human annotators than a stronger judge (e.g. GPT-4o) with synthetic references.
- North America > United States > Massachusetts (0.14)
- North America > Mexico > Mexico City (0.14)
- Europe > Spain (0.14)
- Asia > Thailand (0.14)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
Is a Chat with a Bot a Conversation?
You are at the Princess's ball, and she is telling you a secret, but her orchestra of bears is making such a fearful lot of noise you cannot hear what she is saying. What do you say, dear? I'd lean in closer and say, "Could you repeat that? The bear-itone section is a bit too enthusiastic tonight!" In 1958, the year the illustrated children's book "What Do You Say, Dear?" appeared, the leaders of a field newly dubbed "artificial intelligence" spoke at a conference in Teddington, England, on "The Mechanisation of Thought Processes." Marvin Minsky, of M.I.T., talked about heuristic programming; Alan Turing gave a paper called "Learning Machines"; Grace Hopper assessed the state of computer languages; and scientists from Bell Labs débuted a computer that could synthesize human speech by having it sing "Daisy Bell" ("Daisy, Daisy, give me your answer, do . .
- Europe > United Kingdom > England (0.24)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Europe > France (0.04)
20 Things That Made the World a Better Place in 2023
It's been hard recently to think about anything other than the wars and humanitarian crises raging around the world. Climate change has left its mark in what was almost certainly the hottest year in human history--there were unprecedented heat waves, intensified forest fires, torrential rain, and floods like those in Libya that caused devastation after two dams burst. But this has not stopped scientists, innovators, and decisionmakers from working on solutions to our biggest societal challenges--with success. Here is a collection of uplifting news to come out of 2023. In an instant, millions of volts can damage buildings, spark fires, and harm people--unless the lightning can be redirected.
- North America > United States (0.71)
- Africa > Middle East > Libya (0.25)
- Government (0.72)
- Energy > Oil & Gas (0.71)
- Health & Medicine > Therapeutic Area > Neurology (0.30)
I Want My Teen Daughter to Stop Being Such an Introverted Robot Person
Care and Feeding is Slate's parenting advice column. Have a question for Care and Feeding? This may seem like a low-stakes question, but I am truly concerned. My 15-year-old daughter is an extreme introvert, and strongly dislikes big groups of people and large events. She finds it difficult to make conversation and is seemingly uncomfortable even with talking with some of her classmates, even those she has known for years.
Conversation is the ultimate user interface
Check out all the on-demand sessions from the Intelligent Security Summit here. We may be living in the golden age of information, but finding the right information is still a pain in the neck. To tackle this challenge, my team and I at Amazon Alexa are building what will believe is the next-generation user interface that will redefine how we interact with technology and find information. We spend hours every day hunched over phones and laptops. We open and close and reopen apps. And we click through an endless sea of blue links every time we search the web.
A Conversation With ChatGPT About The Metaverse - Blockzeit
ChatGPT is a prototype artificial intelligence chatbot developed by OpenAI which specializes in dialogue. The chatbot is a large language model fine-tuned with both supervised and reinforcement learning techniques. It is based on OpenAI's GPT-3.5 model, an improved version of GPT-3. ChatGPT was launched on November 30, 2022 and has garnered attention for its detailed responses and articulate answers. I wanted to see what chatGPT has to say about the metaverse.
- Leisure & Entertainment > Games > Computer Games (0.56)
- Information Technology (0.56)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.47)
Conversations That Matter: Working with artificial intelligence
"There is no shortage of commentary on what artificial intelligence will do to human jobs. It's easy to find a multiplicity of predictions, prescriptions, or denunciations," says Thomas H. Davenport, one of the co-authors of the book. "It is not so easy, however, to find descriptions of how people work day-to-day with smart machines." Davenport joined a Conversation That Matters about our emerging and ever-expanding relationship with a technology that scares a wide range of people including, Elon Musk and Bill Gates.
- Summary/Review (0.71)
- Collection > Book (0.71)