Cluster-wise Graph Transformer with Dual-granularity Kernelized Attention 1 Jiayue Zhou
–Neural Information Processing Systems
In the realm of graph learning, there is a category of methods that conceptualize graphs as hierarchical structures, utilizing node clustering to capture broader structural information. While generally effective, these methods often rely on a fixed graph coarsening routine, leading to overly homogeneous cluster representations and loss of node-level information. In this paper, we envision the graph as a network of interconnected node sets without compressing each cluster into a single embedding. To enable effective information transfer among these node sets, we propose the Node-to-Cluster Attention (N2C-Attn) mechanism. N2C-Attn incorporates techniques from Multiple Kernel Learning into the kernelized attention framework, effectively capturing information at both node and cluster levels. We then devise an efficient form for N2C-Attn using the cluster-wise message-passing framework, achieving linear time complexity. We further analyze how N2C-Attn combines bi-level feature maps of queries and keys, demonstrating its capability to merge dual-granularity information. The resulting architecture, Cluster-wise Graph Transformer (Cluster-GT), which uses node clusters as tokens and employs our proposed N2C-Attn module, shows superior performance on various graph-level tasks.
Neural Information Processing Systems
Mar-19-2025, 19:57:22 GMT
- Country:
- Europe (1.00)
- North America > United States
- California > Los Angeles County > Long Beach (0.14)
- Genre:
- Overview (0.67)
- Research Report
- Experimental Study (1.00)
- New Finding (0.93)
- Industry:
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (0.93)
- Statistical Learning (0.68)
- Natural Language (1.00)
- Machine Learning
- Communications (1.00)
- Data Science > Data Mining (1.00)
- Artificial Intelligence
- Information Technology