Improving Expressive Power of Spectral Graph Neural Networks with Eigenvalue Correction

Kangkang Lu,Yanhua Yu,Hao Fei, Xuan Li, Zixuan Yang, Zirui Guo,Meiyu Liang, Mengran Yin,Tat-Seng Chua

AAAI 2024(2024)

引用 0|浏览7
暂无评分
摘要
In recent years, spectral graph neural networks, characterized by polynomial filters, have garnered increasing attention and have achieved remarkable performance in tasks such as node classification. These models typically assume that eigenvalues for the normalized Laplacian matrix are distinct from each other, thus expecting a polynomial filter to have a high fitting ability. However, this paper empirically observes that normalized Laplacian matrices frequently possess repeated eigenvalues. Moreover, we theoretically establish that the number of distinguishable eigenvalues plays a pivotal role in determining the expressive power of spectral graph neural networks. In light of this observation, we propose an eigenvalue correction strategy that can free polynomial filters from the constraints of repeated eigenvalue inputs. Concretely, the proposed eigenvalue correction strategy enhances the uniform distribution of eigenvalues, thus mitigating repeated eigenvalues, and improving the fitting capacity and expressive power of polynomial filters. Extensive experimental results on both synthetic and real-world datasets demonstrate the superiority of our method.
更多
查看译文
关键词
ML: Graph-based Machine Learning,DMKM: Graph Mining, Social Network Analysis & Community
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要