Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Reconsidering the Performance of GAE in Link Prediction

Created by
  • Haebom

Author

Weishuo Ma, Yanbo Wang, Xiyuan Wang, Muhan Zhang

Outline

This paper highlights that the effectiveness of recently published sophisticated learning techniques and model architectures in graph neural networks (GNNs) for link prediction can be exaggerated when compared to older baseline models. To address this, we systematically explore GAEs by applying model-independent techniques used in state-of-the-art methods to graph autoencoders (GAEs) and tuning hyperparameters. We find that well-tuned GAEs perform similarly to recent sophisticated models while offering superior computational efficiency. Specifically, we achieve significant performance gains on datasets with dominant structural information and limited feature data, achieving a state-of-the-art Hits@100 score of 78.41% on the ogbl-ppa dataset. We also analyze the impact of various techniques to elucidate the reasons for their success and suggest future directions. This study highlights the need to update baseline models to more accurately assess the progress of GNNs for link prediction.

Takeaways, Limitations

Takeaways:
We demonstrate that a well-tuned GAE is computationally efficient, achieving performance comparable to state-of-the-art sophisticated GNN models.
Demonstrating the superiority of GAE on datasets rich in structural information and lacking feature data.
Achieves state-of-the-art performance on the ogbl-ppa dataset (Hits@100: 78.41%).
In existing GNN-based link prediction studies, the importance of the baseline model and the need for updating are emphasized.
Limitations:
It cannot be assumed that the proposed method guarantees optimal performance for all types of graph datasets.
Detailed explanation of hyperparameter tuning may be lacking.
Further analysis may be needed to determine the factors contributing to GAE's improved performance.
👍