imposter
Imposter used AI to pose as Marco Rubio and contact foreign ministers
The incident was first revealed in the State Department cable that was dated 3 July and sent to "all diplomatic and consular posts," CBS News reported. The cable stated that a false Signal account was created in mid-June with the display name marco.rubio@state.gov. That account contacted at least five people. "The actor left voicemails on Signal for at least two targeted individuals, and in one instance, sent a text message inviting the individual to communicate on Signal," the cable stated, as reported by CBS. The cable did not identify the individuals that were contacted or what the AI-generated voice of Rubio said in those voicemails.
Training Language Models for Social Deduction with Multi-Agent Reinforcement Learning
Sarkar, Bidipta, Xia, Warren, Liu, C. Karen, Sadigh, Dorsa
Communicating in natural language is a powerful tool in multi-agent settings, as it enables independent agents to share information in partially observable settings and allows zero-shot coordination with humans. However, most prior works are limited as they either rely on training with large amounts of human demonstrations or lack the ability to generate natural and useful communication strategies. In this work, we train language models to have productive discussions about their environment in natural language without any human demonstrations. We decompose the communication problem into listening and speaking. Our key idea is to leverage the agent's goal to predict useful information about the world as a dense reward signal that guides communication. Specifically, we improve a model's listening skills by training them to predict information about the environment based on discussions, and we simultaneously improve a model's speaking skills with multi-agent reinforcement learning by rewarding messages based on their influence on other agents. To investigate the role and necessity of communication in complex social settings, we study an embodied social deduction game based on Among Us, where the key question to answer is the identity of an adversarial imposter. We analyze emergent behaviors due to our technique, such as accusing suspects and providing evidence, and find that it enables strong discussions, doubling the win rates compared to standard RL. We release our code and models at https://socialdeductionllm.github.io/
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- (9 more...)
- Leisure & Entertainment > Games (0.46)
- Education (0.46)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents > Agent Societies (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Imposter.AI: Adversarial Attacks with Hidden Intentions towards Aligned Large Language Models
Liu, Xiao, Li, Liangzhi, Xiang, Tong, Ye, Fuying, Wei, Lu, Li, Wangyue, Garcia, Noa
With the development of large language models (LLMs) like ChatGPT, both their vast applications and potential vulnerabilities have come to the forefront. While developers have integrated multiple safety mechanisms to mitigate their misuse, a risk remains, particularly when models encounter adversarial inputs. This study unveils an attack mechanism that capitalizes on human conversation strategies to extract harmful information from LLMs. We delineate three pivotal strategies: (i) decomposing malicious questions into seemingly innocent sub-questions; (ii) rewriting overtly malicious questions into more covert, benign-sounding ones; (iii) enhancing the harmfulness of responses by prompting models for illustrative examples. Unlike conventional methods that target explicit malicious responses, our approach delves deeper into the nature of the information provided in responses. Through our experiments conducted on GPT-3.5-turbo, GPT-4, and Llama2, our method has demonstrated a marked efficacy compared to conventional attack methods. In summary, this work introduces a novel attack method that outperforms previous approaches, raising an important question: How to discern whether the ultimate intent in a dialogue is malicious?
Gotcha: Real-Time Video Deepfake Detection via Challenge-Response
Mittal, Govind, Hegde, Chinmay, Memon, Nasir
With the rise of AI-enabled Real-Time Deepfakes (RTDFs), the integrity of online video interactions has become a growing concern. RTDFs have now made it feasible to replace an imposter's face with their victim in live video interactions. Such advancement in deepfakes also coaxes detection to rise to the same standard. However, existing deepfake detection techniques are asynchronous and hence ill-suited for RTDFs. To bridge this gap, we propose a challenge-response approach that establishes authenticity in live settings. We focus on talking-head style video interaction and present a taxonomy of challenges that specifically target inherent limitations of RTDF generation pipelines. We evaluate representative examples from the taxonomy by collecting a unique dataset comprising eight challenges, which consistently and visibly degrades the quality of state-of-the-art deepfake generators. These results are corroborated both by humans and a new automated scoring function, leading to 88.6\% and 73.2% AUC, respectively. The findings underscore the promising potential of challenge-response systems for explainable and scalable real-time deepfake detection in practical scenarios.
- Europe > Ukraine (0.14)
- Asia > China (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (5 more...)
SLMIA-SR: Speaker-Level Membership Inference Attacks against Speaker Recognition Systems
Chen, Guangke, Zhang, Yedi, Song, Fu
Membership inference attacks allow adversaries to determine whether a particular example was contained in the model's training dataset. While previous works have confirmed the feasibility of such attacks in various applications, none has focused on speaker recognition (SR), a promising voice-based biometric recognition technique. In this work, we propose SLMIA-SR, the first membership inference attack tailored to SR. In contrast to conventional example-level attack, our attack features speaker-level membership inference, i.e., determining if any voices of a given speaker, either the same as or different from the given inference voices, have been involved in the training of a model. It is particularly useful and practical since the training and inference voices are usually distinct, and it is also meaningful considering the open-set nature of SR, namely, the recognition speakers were often not present in the training data. We utilize intra-similarity and inter-dissimilarity, two training objectives of SR, to characterize the differences between training and non-training speakers and quantify them with two groups of features driven by carefully-established feature engineering to mount the attack. To improve the generalizability of our attack, we propose a novel mixing ratio training strategy to train attack models. To enhance the attack performance, we introduce voice chunk splitting to cope with the limited number of inference voices and propose to train attack models dependent on the number of inference voices. Our attack is versatile and can work in both white-box and black-box scenarios. Additionally, we propose two novel techniques to reduce the number of black-box queries while maintaining the attack performance. Extensive experiments demonstrate the effectiveness of SLMIA-SR.
- North America > United States > California > San Diego County > San Diego (0.04)
- Asia > Singapore (0.04)
- Asia > China (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Pattern Recognition > Speech Recognition (0.61)
To Tell The Truth: Language of Deception and Language Models
Majumder, Bodhisattwa Prasad, Hazra, Sanchaita
Text-based misinformation permeates online discourses, yet evidence of people's ability to discern truth from such deceptive textual content is scarce. We analyze a novel TV game show data where conversations in a high-stake environment between individuals with conflicting objectives result in lies. We investigate the manifestation of potentially verifiable language cues of deception in the presence of objective truth, a distinguishing feature absent in previous text-based deception datasets. We show that there exists a class of detectors (algorithms) that have similar truth detection performance compared to human subjects, even when the former accesses only the language cues while the latter engages in conversations with complete access to all potential sources of cues (language and audio-visual). Our model, built on a large language model, employs a bottleneck framework to learn discernible cues to determine truth, an act of reasoning in which human subjects often perform poorly, even with incentives. Our model detects novel but accurate language cues in many cases where humans failed to detect deception, opening up the possibility of humans collaborating with algorithms and ameliorating their ability to detect the truth.
- North America > Panama (0.04)
- Europe > United Kingdom > England (0.04)
- North America > United States > Tennessee (0.04)
- (6 more...)
- Leisure & Entertainment (1.00)
- Law (0.94)
- Government > Regional Government (0.46)
- Media > News (0.34)
Meta-Learning Framework for End-to-End Imposter Identification in Unseen Speaker Recognition
Chaubey, Ashutosh, Sinha, Sparsh, Ghose, Susmita
Speaker identification systems are deployed in diverse environments, often different from the lab conditions on which they are trained and tested. In this paper, first, we show the problem of generalization using fixed thresholds (computed using EER metric) for imposter identification in unseen speaker recognition and then introduce a robust speaker-specific thresholding technique for better performance. Secondly, inspired by the recent use of meta-learning techniques in speaker verification, we propose an end-to-end meta-learning framework for imposter detection which decouples the problem of imposter detection from unseen speaker identification. Thus, unlike most prior works that use some heuristics to detect imposters, the proposed network learns to detect imposters by leveraging the utterances of the enrolled speakers. Furthermore, we show the efficacy of the proposed techniques on VoxCeleb1, VCTK and the FFSVC 2022 datasets, beating the baselines by up to 10%.
Can YOU guess the odd one out? Google challenges you to find AI 'imposters' hidden among atworks
From an image of Pope Francis wearing a puffer jacket to a snap of Donald Trump being arrested in New York City, several AI-generated images have fooled onlookers in recent months. Now, Google is putting our ability to spot AI-generated content to the test in a new game, called Odd One Out. The game shows players four artworks and tasks them with spotting the one that has been generated by AI. While it might sound easy, the game is fiendishly difficult, with the AI creating impressively convincing artworks. 'Can you spot the odd one out?
The Video Game AOC Keeps Streaming Is Actually a Good Metaphor for Our Politics
It's a fight in the realm of public opinion, which has been poisoned by bad-faith actors whose lies are threatening to overcome the truth. That sure sounds like the political environment we inhabit, but I'm talking about Among Us, the hugely popular indie video game where disguised killers stalk a crew of astronauts as they try to repair their broken ship and escape with their lives. Among Us has become a favorite of Rep. Alexandria Ocasio-Cortez, who has now live-streamed herself playing the game twice. The most recent stream, the day after Thanksgiving, brought in an audience of more than 2 million viewers. Her youthful, relatable communication style and commitment to bringing her message into the digital realms where many conservatives have been thriving for years makes it a savvy choice.
- Information Technology > Communications > Social Media (0.89)
- Information Technology > Artificial Intelligence > Games (0.61)
Among Us is the ultimate party game of the Covid era
There are 10 crew members trapped on a spacecraft, carrying out menial tasks to maintain vital systems, but at least one of them is an imposter who wants to sabotage their work and if possible, murder them. What sounds like the premise of a particularly bleak science-fiction movie is in fact the set-up of one of the most popular video games of the year. Developed by a three-person team at InnerSloth and launched to virtual obscurity in 2018, Among Us has suddenly become one of the biggest games on PC and mobile, attracting more than 85m players in the last six months. It's so successful, InnerSloth recently abandoned plans to work on a sequel, instead piling their resources into the original. No one, it seems, is more surprised about the success of this game than its creators.