Style Attack Disguise: When Fonts Become a Camouflage for Adversarial Intent

Zhang, Yangshijie, Wang, Xinda, Liu, Jialin, Wang, Wenqiang, Ma, Zhicong, Jia, Xingxing

arXiv.org Artificial Intelligence 

ABSTRACT With social media growth, users employ stylistic fonts and font-like emoji to express individuality, creating visually appealing text that remains human-readable. However, these fonts introduce hidden vulnerabilities in NLP models: while humans easily read stylistic text, models process these characters as distinct tokens, causing interference. We identify this human-model perception gap and propose a style-based attack, Style Attack Disguise (SAD). Experiments on sentiment classification and machine translation across traditional models, LLMs, and commercial services demonstrate SAD's strong attack performance. We also show SAD's potential threats to multimodal tasks including text-to-image and text-to-speech generation.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found