Back to Search Start Over

Overcoming Pitfalls in Graph Contrastive Learning Evaluation: Toward Comprehensive Benchmarks

Authors :
Ma, Qian
Chi, Hongliang
Zhang, Hengrui
Liu, Kay
Zhang, Zhiwei
Cheng, Lu
Wang, Suhang
Yu, Philip S.
Ma, Yao
Publication Year :
2024

Abstract

The rise of self-supervised learning, which operates without the need for labeled data, has garnered significant interest within the graph learning community. This enthusiasm has led to the development of numerous Graph Contrastive Learning (GCL) techniques, all aiming to create a versatile graph encoder that leverages the wealth of unlabeled data for various downstream tasks. However, the current evaluation standards for GCL approaches are flawed due to the need for extensive hyper-parameter tuning during pre-training and the reliance on a single downstream task for assessment. These flaws can skew the evaluation away from the intended goals, potentially leading to misleading conclusions. In our paper, we thoroughly examine these shortcomings and offer fresh perspectives on how GCL methods are affected by hyper-parameter choices and the choice of downstream tasks for their evaluation. Additionally, we introduce an enhanced evaluation framework designed to more accurately gauge the effectiveness, consistency, and overall capability of GCL methods.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2402.15680
Document Type :
Working Paper