Goto

Collaborating Authors

 conversation


Can YOU see him? Take the test to see if you can spot Jesus in objects thanks to unusual brain phenomenon

Daily Mail - Science & tech

With his flowing locks, long beard, and worn robes, Jesus is one of the most instantly recognisable figures in the Western world. So it comes as no surprise that his face is also regularly spotted in inanimate objects. This is due to'face pareidolia' - a common brain phenomenon in which a person sees faces in random images or patterns. 'Sometimes we see faces that aren't really there,' explained Robin Kramer, Senior Lecturer in the School of Psychology, at University of Lincoln, in an article for The Conversation. 'You may be looking at the front of a car or a burnt piece of toast when you notice a face-like pattern. 'This is called face pareidolia and is a mistake made by the brain's face detection system.'


Can one size fit all?: Measuring Failure in Multi-Document Summarization Domain Transfer

DeLucia, Alexandra, Dredze, Mark

arXiv.org Artificial Intelligence

Abstractive multi-document summarization (MDS) is the task of automatically summarizing information in multiple documents, from news articles to conversations with multiple speakers. The training approaches for current MDS models can be grouped into four approaches: end-to-end with special pre-training ("direct"), chunk-then-summarize, extract-then-summarize, and inference with GPT-style models. In this work, we evaluate MDS models across training approaches, domains, and dimensions (reference similarity, quality, and factuality), to analyze how and why models trained on one domain can fail to summarize documents from another (News, Science, and Conversation) in the zero-shot domain transfer setting. We define domain-transfer "failure" as a decrease in factuality, higher deviation from the target, and a general decrease in summary quality. In addition to exploring domain transfer for MDS models, we examine potential issues with applying popular summarization metrics out-of-the-box.


Personalized Attacks of Social Engineering in Multi-turn Conversations -- LLM Agents for Simulation and Detection

Kumarage, Tharindu, Johnson, Cameron, Adams, Jadie, Ai, Lin, Kirchner, Matthias, Hoogs, Anthony, Garland, Joshua, Hirschberg, Julia, Basharat, Arslan, Liu, Huan

arXiv.org Artificial Intelligence

The rapid advancement of conversational agents, particularly chatbots powered by Large Language Models (LLMs), poses a significant risk of social engineering (SE) attacks on social media platforms. SE detection in multi-turn, chat-based interactions is considerably more complex than single-instance detection due to the dynamic nature of these conversations. A critical factor in mitigating this threat is understanding the mechanisms through which SE attacks operate, specifically how attackers exploit vulnerabilities and how victims' personality traits contribute to their susceptibility. In this work, we propose an LLM-agentic framework, SE-VSim, to simulate SE attack mechanisms by generating multi-turn conversations. We model victim agents with varying personality traits to assess how psychological profiles influence susceptibility to manipulation. Using a dataset of over 1000 simulated conversations, we examine attack scenarios in which adversaries, posing as recruiters, funding agencies, and journalists, attempt to extract sensitive information. Based on this analysis, we present a proof of concept, SE-OmniGuard, to offer personalized protection to users by leveraging prior knowledge of the victims personality, evaluating attack strategies, and monitoring information exchanges in conversations to identify potential SE attempts.


No Free Labels: Limitations of LLM-as-a-Judge Without Human Grounding

Krumdick, Michael, Lovering, Charles, Reddy, Varshini, Ebner, Seth, Tanner, Chris

arXiv.org Artificial Intelligence

LLM-as-a-Judge is a framework that uses an LLM (large language model) to evaluate the quality of natural language text - typically text that is also generated by an LLM. This framework holds great promise due to its relative low-cost, ease of use, and strong correlations with human stylistic preferences. However, LLM Judges have been shown to exhibit biases that can distort their judgments. We evaluate how well LLM Judges can grade whether a given response to a conversational question is correct, an ability crucial to soundly estimating the overall response quality. To do so, we create and publicly release a human-annotated dataset with labels of correctness for 1,200 LLM responses. We source questions from a combination of existing datasets and a novel, challenging benchmark (BFF-Bench) created for this analysis. We demonstrate a strong connection between an LLM's ability to correctly answer a question and grade responses to that question. Although aggregate level statistics might imply a judge has high agreement with human annotators, it will struggle on the subset of questions it could not answer. To address this issue, we recommend a simple solution: provide the judge with a correct, human-written reference answer. We perform an in-depth analysis on how reference quality can affect the performance of an LLM Judge. We show that providing a weaker judge (e.g. Qwen 2.5 7B) with higher quality references reaches better agreement with human annotators than a stronger judge (e.g. GPT-4o) with synthetic references.


Is a Chat with a Bot a Conversation?

The New Yorker

You are at the Princess's ball, and she is telling you a secret, but her orchestra of bears is making such a fearful lot of noise you cannot hear what she is saying. What do you say, dear? I'd lean in closer and say, "Could you repeat that? The bear-itone section is a bit too enthusiastic tonight!" In 1958, the year the illustrated children's book "What Do You Say, Dear?" appeared, the leaders of a field newly dubbed "artificial intelligence" spoke at a conference in Teddington, England, on "The Mechanisation of Thought Processes." Marvin Minsky, of M.I.T., talked about heuristic programming; Alan Turing gave a paper called "Learning Machines"; Grace Hopper assessed the state of computer languages; and scientists from Bell Labs débuted a computer that could synthesize human speech by having it sing "Daisy Bell" ("Daisy, Daisy, give me your answer, do . .


20 Things That Made the World a Better Place in 2023

WIRED

It's been hard recently to think about anything other than the wars and humanitarian crises raging around the world. Climate change has left its mark in what was almost certainly the hottest year in human history--there were unprecedented heat waves, intensified forest fires, torrential rain, and floods like those in Libya that caused devastation after two dams burst. But this has not stopped scientists, innovators, and decisionmakers from working on solutions to our biggest societal challenges--with success. Here is a collection of uplifting news to come out of 2023. In an instant, millions of volts can damage buildings, spark fires, and harm people--unless the lightning can be redirected.


I Want My Teen Daughter to Stop Being Such an Introverted Robot Person

Slate

Care and Feeding is Slate's parenting advice column. Have a question for Care and Feeding? This may seem like a low-stakes question, but I am truly concerned. My 15-year-old daughter is an extreme introvert, and strongly dislikes big groups of people and large events. She finds it difficult to make conversation and is seemingly uncomfortable even with talking with some of her classmates, even those she has known for years.

  Country: North America > United States > New York (0.04)
  Genre: Personal > Human Interest (0.40)
  Industry: Education (0.47)

Conversation is the ultimate user interface

#artificialintelligence

Check out all the on-demand sessions from the Intelligent Security Summit here. We may be living in the golden age of information, but finding the right information is still a pain in the neck. To tackle this challenge, my team and I at Amazon Alexa are building what will believe is the next-generation user interface that will redefine how we interact with technology and find information. We spend hours every day hunched over phones and laptops. We open and close and reopen apps. And we click through an endless sea of blue links every time we search the web.


A Conversation With ChatGPT About The Metaverse - Blockzeit

#artificialintelligence

ChatGPT is a prototype artificial intelligence chatbot developed by OpenAI which specializes in dialogue. The chatbot is a large language model fine-tuned with both supervised and reinforcement learning techniques. It is based on OpenAI's GPT-3.5 model, an improved version of GPT-3. ChatGPT was launched on November 30, 2022 and has garnered attention for its detailed responses and articulate answers. I wanted to see what chatGPT has to say about the metaverse.


Conversations That Matter: Working with artificial intelligence

#artificialintelligence

"There is no shortage of commentary on what artificial intelligence will do to human jobs. It's easy to find a multiplicity of predictions, prescriptions, or denunciations," says Thomas H. Davenport, one of the co-authors of the book. "It is not so easy, however, to find descriptions of how people work day-to-day with smart machines." Davenport joined a Conversation That Matters about our emerging and ever-expanding relationship with a technology that scares a wide range of people including, Elon Musk and Bill Gates.

  Country: North America > Canada (0.40)
  Genre:
  Industry: Energy > Oil & Gas > Upstream (0.40)