A Split-Window Transformer for Multi-Model Sequence Spammer Detection using Multi-Model Variational Autoencoder
Yang, Zhou, Pang, Yucai, Yin, Hongbo, Xiao, Yunpeng
–arXiv.org Artificial Intelligence
This paper introduces a new Transformer, called MS$^2$Dformer, that can be used as a generalized backbone for multi-modal sequence spammer detection. Spammer detection is a complex multi-modal task, thus the challenges of applying Transformer are two-fold. Firstly, complex multi-modal noisy information about users can interfere with feature mining. Secondly, the long sequence of users' historical behaviors also puts a huge GPU memory pressure on the attention computation. To solve these problems, we first design a user behavior Tokenization algorithm based on the multi-modal variational autoencoder (MVAE). Subsequently, a hierarchical split-window multi-head attention (SW/W-MHA) mechanism is proposed. The split-window strategy transforms the ultra-long sequences hierarchically into a combination of intra-window short-term and inter-window overall attention. Pre-trained on the public datasets, MS$^2$Dformer's performance far exceeds the previous state of the art. The experiments demonstrate MS$^2$Dformer's ability to act as a backbone.
arXiv.org Artificial Intelligence
Feb-23-2025
- Country:
- Asia > China
- Beijing > Beijing (0.04)
- Chongqing Province > Chongqing (0.05)
- Heilongjiang Province > Harbin (0.04)
- Henan Province > Zhengzhou (0.04)
- Sichuan Province > Chengdu (0.04)
- North America > Canada
- Alberta > Census Division No. 13 > Woodlands County (0.04)
- Asia > China
- Genre:
- Research Report (0.50)
- Technology: