G-OSR: A Comprehensive Benchmark for Graph Open-Set Recognition
Dong, Yicong, He, Rundong, Chen, Guangyao, Zhang, Wentao, Han, Zhongyi, Shi, Jieming, Yin, Yilong
–arXiv.org Artificial Intelligence
--Graph Neural Networks (GNNs) have achieved significant success in machine learning, with wide applications in social networks, bioinformatics, knowledge graphs, and other fields. Most research assumes ideal closed-set environments. However, in real-world open-set environments, graph learning models face challenges in robustness and reliability due to unseen classes. This highlights the need for Graph Open-Set Recognition (GOSR) methods to address these issues and ensure effective GNN application in practical scenarios. Research in GOSR is in its early stages, with a lack of a comprehensive benchmark spanning diverse tasks and datasets to evaluate methods. Moreover, traditional methods, Graph Out-of-Distribution Detection (GOODD), GOSR, and Graph Anomaly Detection (GAD) have mostly evolved in isolation, with little exploration of their interconnections or potential applications to GOSR. T o fill these gaps, we introduce G-OSR, a comprehensive benchmark for evaluating GOSR methods at both the node and graph levels, using datasets from multiple domains to ensure fair and standardized comparisons of effectiveness and efficiency across traditional, GOODD, GOSR, and GAD methods. The results offer critical insights into the generalizability and limitations of current GOSR methods and provide valuable resources for advancing research in this field through systematic analysis of diverse approaches. RAPH learning, as a significant research direction in machine learning, has been widely applied in social network analysis, recommendation systems, bioinformatics, knowledge graphs, traffic planning, and the fields of chemistry and materials science [1]. Graph Neural Networks (GNNs) have demonstrated superior performance in various node classification and graph classification tasks [2]. These methods typically follow a closed-set setting, which assumes that all test classes are among the seen classes accessible during training [3]. However, in real-world scenarios, due to undersampling, out-of-distribution, or anomalous samples, it is highly likely to encounter samples belonging to novel unseen classes, which can significantly impact the safety and robustness of models [4], as illustrated in Figure 1. Guangyao Chen is with Cornell University, Ithaca, NY, USA. Wentao Zhang is with Peking University, Beijing, China. Zhongyi Han is with King Abdullah University of Science and Technology, Thuwal, Saudi Arabia. Rundong He and Yilong Yin are the corresponding authors. Closed-set classification cannot identify unseen classes, while open-set recognition can identify unseen classes and classify nodes belonging to seen classes.
arXiv.org Artificial Intelligence
Mar-1-2025
- Country:
- Asia
- China > Beijing
- Beijing (0.24)
- Middle East > Saudi Arabia
- Mecca Province > Thuwal (0.24)
- China > Beijing
- North America > United States
- New York > Tompkins County > Ithaca (0.24)
- Asia
- Genre:
- Overview (0.87)
- Research Report (1.00)
- Industry:
- Health & Medicine > Pharmaceuticals & Biotechnology (1.00)
- Information Technology > Services (0.86)
- Technology: