Beyond Over-smoothing: Uncovering the Trainability Challenges in Deep Graph Neural Networks
Peng, Jie, Lei, Runlin, Wei, Zhewei
–arXiv.org Artificial Intelligence
The drastic performance degradation of Graph Neural Networks (GNNs) as the depth of the graph propagation layers exceeds 8-10 is widely attributed to a phenomenon of Over-smoothing. Although recent research suggests that Over-smoothing may not be the dominant reason for such a performance degradation, they have not provided rigorous analysis from a theoretical view, which warrants further investigation. In this paper, we systematically analyze the real dominant problem in deep GNNs and identify the issues that these GNNs towards addressing Over-smoothing essentially work on via empirical experiments and theoretical gradient analysis. We theoretically prove that the difficult training problem of deep MLPs is actually the main challenge, and various existing methods that supposedly tackle Over-smoothing actually improve the trainability of MLPs, which is the main reason for their performance gains. Our further investigation into trainability issues reveals that properly constrained smaller upper bounds of gradient flow notably enhance the trainability of GNNs. Experimental results on diverse datasets demonstrate consistency between our theoretical findings and empirical evidence. Our analysis provides new insights in constructing deep graph models.
arXiv.org Artificial Intelligence
Aug-7-2024
- Country:
- Asia
- China
- Beijing > Beijing (0.05)
- Guangdong Province > Guangzhou (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- China
- North America > United States
- Idaho > Ada County
- Boise (0.05)
- New York > New York County
- New York City (0.04)
- Idaho > Ada County
- Asia
- Genre:
- Research Report > New Finding (0.86)
- Technology: