Local Virtual Nodes for Alleviating Over-Squashing in Graph Neural Networks

Karabulut, Tuğrul Hasan, Baytaş, İnci M.

arXiv.org Artificial Intelligence 

--Over-squashing is a challenge in training graph neural networks for tasks involving long-range dependencies. In such tasks, a GNN's receptive field should be large enough to enable communication between distant nodes. However, gathering information from a wide range of neighborhoods and squashing its content into fixed-size node representations makes message-passing vulnerable to bottlenecks. Graph rewiring and adding virtual nodes are commonly studied remedies that create additional pathways around bottlenecks to mitigate over-squashing. However, these techniques alter the input graph's global topology and disrupt the domain knowledge encoded in the original graph structure, both of which could be essential to specific tasks and domains. This study presents Local Virtual Nodes (L VN) with trainable embeddings to alleviate the effects of over-squashing without significantly corrupting the global structure of the input graph. The position of the L VNs is determined by the node centrality, which indicates the existence of potential bottlenecks. Thus, the proposed approach aims to improve the connectivity in the regions with likely bottlenecks. Furthermore, trainable L VN embeddings shared across selected central regions facilitate communication between distant nodes without adding more layers. Extensive experiments on benchmark datasets demonstrate that L VNs can enhance structural connectivity and significantly improve performance on graph and node classification tasks. The code can be found at https://github.com/ALLab-Boun/L RAPH Neural Networks (GNNs) have become a standard for representation learning in tasks involving non-Euclidean data. They can handle arbitrary graph topologies without any prior assumptions about their structure. Therefore, the GNNs are integral components in applications of social networks [1], traffic networks [2], and molecular graphs [3]. The flexibility of GNNs and their applicability to various domains mainly stem from the message-passing paradigm, an efficient operation that can handle diverse graph structures at a massive scale [4], [5].