I. Introduction
Self-supervised learning provides a promising learning paradigm without relying on high-cost label information for many research fields such as computer vision [1]–[3], natural language processing [4]–[7], speech recognition [8]–[10], and recommender systems [11]–[13]. Contrastive-based methods have a prominent place among the landscape of self-supervised learning methods [14]–[17]. Contrastive learning leverages the inherent structure and relationships within unlabeled data to train encoder networks [18]–[20]. The core idea behind contrastive learning is to map positives (i.e., positive samples) closer in the embedding space while pushing negatives (i.e., negative samples) apart. This process encourages the encoder network to capture intricate patterns, semantic relationships, and underlying structures present in the data, making it particularly adept at learning useful embeddings from diverse and complex datasets. Recently, researchers have explored the graph contrastive learning (GCL) framework for self-supervised learning on graphs [21]–[25]. On benchmark datasets, the state-of-the-art GCL models have demonstrated competitive performance against supervised learning models, e.g., graph convolutional network (GCN) [26], in various graph-related tasks such as node classification [27]–[31], graph classification [32], [33], and link prediction [34]–[37].