Topology-Aware Dynamic Reweighting for Distribution Shifts on Graph
Zheng, Weihuang, Liu, Jiashuo, Li, Jiaxing, Wu, Jiayun, Cui, Peng, Kong, Youyong
–arXiv.org Artificial Intelligence
Graph Neural Networks (GNNs) have been widely used in node classification tasks, such as advertising recommendation [15], social network anomaly detection [34], etc. However, these GNN models typically assume that the training and test graph data are drawn from the same distribution, which does not always hold in practice. In real-world graph data, sample selection bias [8, 12] as well as graph construction techniques [27, 43] often brings distribution shifts between training nodes and test nodes. For instance, In WebKB [26] datasets, web pages (nodes) and categories (labels) are heavily affected by the university they originate from, leading to distribution shifts among nodes drawn from different universities. Therefore, in order to enhance the practical validity of GNNs, it is of paramount importance to deal with distribution shifts on graph data. To address the distribution shift problem in node classification, recent works [18, 36, 32, 37, 23] borrow the idea of invariant learning methods from the literature of out-of-distribution (OOD) generalization and adopt them on graph-structured data. Invariant learning [1, 19] stems from the causal inference literature, and now becomes one of the key approaches to solving OOD problems on graphs. The core concept is to identify invariant features with stable prediction mechanisms across different environments, thereby mitigating performance degradation under distribution shifts. And most of the works in this line directly apply existing invariant learning algorithms to graph-level classification tasks (major) [18, 32, 23, 41] and node classification tasks (minor) [36, 38].
arXiv.org Artificial Intelligence
Jun-3-2024
- Country:
- Europe > Netherlands (0.14)
- North America > United States (0.14)
- Genre:
- Research Report > New Finding (0.68)
- Technology: