Goto

Collaborating Authors

 clare


OpenFact at CheckThat! 2024: Combining Multiple Attack Methods for Effective Adversarial Text Generation

Lewoniewski, Włodzimierz, Stolarski, Piotr, Stróżyna, Milena, Lewańska, Elzbieta, Wojewoda, Aleksandra, Księżniak, Ewelina, Sawiński, Marcin

arXiv.org Artificial Intelligence

This paper presents the experiments and results for the CheckThat! Lab at CLEF 2024 Task 6: Robustness of Credibility Assessment with Adversarial Examples (InCrediblAE). The primary objective of this task was to generate adversarial examples in five problem domains in order to evaluate the robustness of widely used text classification methods (fine-tuned BERT, BiLSTM, and RoBERTa) when applied to credibility assessment issues. This study explores the application of ensemble learning to enhance adversarial attacks on natural language processing (NLP) models. We systematically tested and refined several adversarial attack methods, including BERT-Attack, Genetic algorithms, TextFooler, and CLARE, on five datasets across various misinformation tasks. By developing modified versions of BERT-Attack and hybrid methods, we achieved significant improvements in attack effectiveness. Our results demonstrate the potential of modification and combining multiple methods to create more sophisticated and effective adversarial attack strategies, contributing to the development of more robust and secure systems.


Saliency Attention and Semantic Similarity-Driven Adversarial Perturbation

Waghela, Hetvi, Sen, Jaydip, Rakshit, Sneha

arXiv.org Artificial Intelligence

In this paper, we introduce an enhanced textual adversarial attack method, known as Saliency Attention and Semantic Similarity driven adversarial Perturbation (SASSP). The proposed scheme is designed to improve the effectiveness of contextual perturbations by integrating saliency, attention, and semantic similarity. Traditional adversarial attack methods often struggle to maintain semantic consistency and coherence while effectively deceiving target models. Our proposed approach addresses these challenges by incorporating a three-pronged strategy for word selection and perturbation. First, we utilize a saliency-based word selection to prioritize words for modification based on their importance to the model's prediction. Second, attention mechanisms are employed to focus perturbations on contextually significant words, enhancing the attack's efficacy. Finally, an advanced semantic similarity-checking method is employed that includes embedding-based similarity and paraphrase detection. By leveraging models like Sentence-BERT for embedding similarity and fine-tuned paraphrase detection models from the Sentence Transformers library, the scheme ensures that the perturbed text remains contextually appropriate and semantically consistent with the original. Empirical evaluations demonstrate that SASSP generates adversarial examples that not only maintain high semantic fidelity but also effectively deceive state-of-the-art natural language processing models. Moreover, in comparison to the original scheme of contextual perturbation CLARE, SASSP has yielded a higher attack success rate and lower word perturbation rate.


Realistic Continual Learning Approach using Pre-trained Models

Nasri, Nadia, Gutiérrez-Álvarez, Carlos, Lafuente-Arroyo, Sergio, Maldonado-Bascón, Saturnino, López-Sastre, Roberto J.

arXiv.org Artificial Intelligence

Continual learning (CL) is crucial for evaluating adaptability in learning solutions to retain knowledge. Our research addresses the challenge of catastrophic forgetting, where models lose proficiency in previously learned tasks as they acquire new ones. While numerous solutions have been proposed, existing experimental setups often rely on idealized class-incremental learning scenarios. We introduce Realistic Continual Learning (RealCL), a novel CL paradigm where class distributions across tasks are random, departing from structured setups. We also present CLARE (Continual Learning Approach with pRE-trained models for RealCL scenarios), a pre-trained model-based solution designed to integrate new knowledge while preserving past learning. Our contributions include pioneering RealCL as a generalization of traditional CL setups, proposing CLARE as an adaptable approach for RealCL tasks, and conducting extensive experiments demonstrating its effectiveness across various RealCL scenarios. Notably, CLARE outperforms existing models on RealCL benchmarks, highlighting its versatility and robustness in unpredictable learning environments.


Blank review – author held hostage by AI as near-future thriller enters Misery territory

The Guardian

In what has the distinctively zoned-out vibe of another lockdown-born project, Natalie Kennedy's sci-fi psychological thriller sees Clare Rivers (Rachel Shelley), an author with writer's block, sign up for a deluxe writing retreat operated entirely by AI. Sealed hermetically into her unit by a virus that corrupts the system, she can't leave until she has produced a book, making Blank play out like Misery and Ex Machina spliced. Taking place in a near future where writing is all holographic word processors and genial AI assistants rather than tattered notebooks and half-eaten Twixes, the profession seems to have moved on. Or perhaps not: Clare's blockage is aggravated by being locked in with only a malfunctioning amnesiac android called Rita (Heida Reed) for company. Reset every day and refusing to open the external doors until Clare has delivered the goods, in the face of the writer's exasperation Rita can only passive-aggressively reel off: "You seem distressed. Maybe you should have a lie down."


Cross-Modal Attribute Insertions for Assessing the Robustness of Vision-and-Language Learning

Ramshetty, Shivaen, Verma, Gaurav, Kumar, Srijan

arXiv.org Artificial Intelligence

The robustness of multimodal deep learning models to realistic changes in the input text is critical for their applicability to important tasks such as text-to-image retrieval and cross-modal entailment. To measure robustness, several existing approaches edit the text data, but do so without leveraging the cross-modal information present in multimodal data. Information from the visual modality, such as color, size, and shape, provide additional attributes that users can include in their inputs. Thus, we propose cross-modal attribute insertions as a realistic perturbation strategy for vision-and-language data that inserts visual attributes of the objects in the image into the corresponding text (e.g., "girl on a chair" to "little girl on a wooden chair"). Our proposed approach for cross-modal attribute insertions is modular, controllable, and task-agnostic. We find that augmenting input text using cross-modal insertions causes state-of-the-art approaches for text-to-image retrieval and cross-modal entailment to perform poorly, resulting in relative drops of 15% in MRR and 20% in $F_1$ score, respectively. Crowd-sourced annotations demonstrate that cross-modal insertions lead to higher quality augmentations for multimodal data than augmentations using text-only data, and are equivalent in quality to original examples. We release the code to encourage robustness evaluations of deep vision-and-language models: https://github.com/claws-lab/multimodal-robustness-xmai.


7 TV shows you need to watch in April 2023

#artificialintelligence

April promises to be an exciting month for streaming TV. Popular shows are returning this month for second seasons, including Apple TV's Schmigadoon! Then there's HBO's Somebody Somewhere, which is back for season 2 on April 23. But there are exciting new shows premiering this month as well that are worth checking out. Amazon Prime Video, for example, has an interesting gender-reversed version of David Cronenberg's 1988 movie Dead Ringers, starring Rachel Weisz in the role previously portrayed by Jeremy Irons.


CLARE: Conservative Model-Based Reward Learning for Offline Inverse Reinforcement Learning

Yue, Sheng, Wang, Guanbo, Shao, Wei, Zhang, Zhaofeng, Lin, Sen, Ren, Ju, Zhang, Junshan

arXiv.org Artificial Intelligence

This work aims to tackle a major challenge in offline Inverse Reinforcement Learning (IRL), namely the reward extrapolation error, where the learned reward function may fail to explain the task correctly and misguide the agent in unseen environments due to the intrinsic covariate shift. Leveraging both expert data and lower-quality diverse data, we devise a principled algorithm (namely CLARE) that solves offline IRL efficiently via integrating "conservatism" into a learned reward function and utilizing an estimated dynamics model. Our theoretical analysis provides an upper bound on the return gap between the learned policy and the expert policy, based on which we characterize the impact of covariate shift by examining subtle two-tier tradeoffs between the "exploitation" (on both expert and diverse data) and "exploration" (on the estimated dynamics model). We show that CLARE can provably alleviate the reward extrapolation error by striking the right "exploitation-exploration" balance therein. Extensive experiments corroborate the significant performance gains of CLARE over existing state-of-the-art algorithms on MuJoCo continuous control tasks (especially with a small offline dataset), and the learned reward is highly instructive for further learning (source code). The primary objective of Inverse Reinforcement Learning (IRL) is to learn a reward function from demonstrations (Arora & Doshi, 2021; Russell, 1998). In general, conventional IRL methods rely on extensive online trials and errors that can be costly or require a fully known transition model (Abbeel & Ng, 2004; Ratliff et al., 2006; Ziebart et al., 2008; Syed & Schapire, 2007; Boularias et al., 2011; Osa et al., 2018), struggling to scale in many real-world applications. To tackle this problem, this paper studies offline IRL, with focus on learning from a previously collected dataset without online interaction with the environment.


A Context-Aware Approach for Textual Adversarial Attack through Probability Difference Guided Beam Search

Liu, Huijun, Yu, Jie, Li, Shasha, Ma, Jun, Ji, Bin

arXiv.org Artificial Intelligence

Textual adversarial attacks expose the vulnerabilities of text classifiers and can be used to improve their robustness. Existing context-aware methods solely consider the gold label probability and use the greedy search when searching an attack path, often limiting the attack efficiency. To tackle these issues, we propose PDBS, a context-aware textual adversarial attack model using Probability Difference guided Beam Search. The probability difference is an overall consideration of all class label probabilities, and PDBS uses it to guide the selection of attack paths. In addition, PDBS uses the beam search to find a successful attack path, thus avoiding suffering from limited search space. Extensive experiments and human evaluation demonstrate that PDBS outperforms previous best models in a series of evaluation metrics, especially bringing up to a +19.5% attack success rate. Ablation studies and qualitative analyses further confirm the efficiency of PDBS.


Two hands are better than one

Robohub

What are you doing right now other than scrolling through this article? Do you have a cup of coffee in one hand, your phone in the other? Maybe your right hand is using your laptop mouse and your left hand is holding a snack. Have you ever thought about how often we are using both of our hands? Having two healthy human hands allows us to carry too many grocery bags in one hand and unlock our apartment door with the other, and perform complex bimanual coordination like playing Moonlight Sonata by Beethoven on the piano (well, maybe not all of us can do that). Having two hands also allows us to do some of the most simple tasks in our daily lives, like holding a jar of peanut butter and unscrewing the lid, or putting our hair up in a ponytail.


Privacy expert Clare Garvie explains why your face is already in a criminal lineup

#artificialintelligence

Biometric surveillance is coming for you, even if you have'nothing to hide' Clare Garvie is a Senior Associate at Georgetown University's Center on Privacy and Technology, where she has dedicated her work to studying law enforcement's use of face recognition technology on the American public. She is considered the foremost expert on face recognition technology; last year she testified in front of Congress. She writes extensively on its use in law enforcement investigations. As well, she brings to light the worrying ways the technology disrupts privacy, circumvents judicial norms and legal precedents, and promotes chilling effects on free speech and civil liberties. All of this happens under a veil of secrecy, without public consent and largely outside of the purview of American lawmakers. Garvie's research spotlights the ways these technologies are disproportionately used on Black and Brown communities and the failures of face recognition algorithms when deployed on people of color and women. The technology's efficacy, already cause for concern, is further problematized by law enforcement's cavalier practices.