Token-Modification Adversarial Attacks for Natural Language Processing: A Survey
Roth, Tom, Gao, Yansong, Abuadbba, Alsharif, Nepal, Surya, Liu, Wei
–arXiv.org Artificial Intelligence
Many adversarial attacks target natural language processing systems, most of which succeed through modifying the individual tokens of a document. Despite the apparent uniqueness of each of these attacks, fundamentally they are simply a distinct configuration of four components: a goal function, allowable transformations, a search method, and constraints. In this survey, we systematically present the different components used throughout the literature, using an attack-independent framework which allows for easy comparison and categorisation of components. Our work aims to serve as a comprehensive guide for newcomers to the field and to spark targeted research into refining the individual attack components.
arXiv.org Artificial Intelligence
Jan-7-2024
- Country:
- Asia (1.00)
- Europe (1.00)
- North America > United States
- California > San Francisco County
- San Francisco (0.14)
- Maryland > Baltimore (0.14)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- California > San Francisco County
- Genre:
- Overview (1.00)
- Industry:
- Government (1.00)
- Information Technology > Security & Privacy (1.00)
- Technology: