Learning Advanced Self-Attention for Linear Transformers in the Singular Value Domain
Wi, Hyowon, Choi, Jeongwhan, Park, Noseong
–arXiv.org Artificial Intelligence
Transformers have demonstrated remarkable performance across diverse domains. The key component of Transformers is self-attention, which learns the relationship between any two tokens in the input sequence. Recent studies have revealed that the self-attention can be understood as a normalized adjacency matrix of a graph. Notably, from the perspective of graph signal processing (GSP), the self-attention can be equivalently defined as a simple graph filter, applying GSP using the value vector as the signal. However, the self-attention is a graph filter defined with only the first order of the polynomial matrix, and acts as a low-pass filter preventing the effective leverage of various frequency information. Consequently, existing self-attention mechanisms are designed in a rather simplified manner. Therefore, we propose a novel method, called A ttentive G raph F ilter (AGF), interpreting the self-attention as learning the graph filter in the singular value domain from the perspective of graph signal processing for directed graphs with the linear complexity w.r.t. the input length n, i.e., O p nd
arXiv.org Artificial Intelligence
May-14-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- North America > United States
- California (0.04)
- South America > Chile
- Asia > Middle East
- Genre:
- Research Report (1.00)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (0.46)
- Statistical Learning (0.69)
- Natural Language (1.00)
- Vision (0.93)
- Machine Learning
- Data Science (0.68)
- Artificial Intelligence
- Information Technology