Introduction
Many systems in nature and society, from the World Wide Web to social and biological networks, can be aptly described as complex networks or graphs [1], [2], [3]. In these networks, nodes or vertices represent entities, while the interactions among these entities are depicted as edges or links [1]. Complex networks, serving as abstract models for understanding real-world systems, exhibit characteristics such as self-organization, self-similarity, small-world properties, and scale-free nature. Link prediction within these networks is a critical task that involves uncovering hidden or emerging links between nodes. This includes not only identifying unknown links that already exist but have not yet been detected in the network but also forecasting future links that, although not currently present, are likely or ought to exist in the foreseeable future [4].
In practical applications, link prediction plays a pivotal role in forecasting potential new connections within a network, thereby supporting and enhancing decision-making processes. For instance, in the field of biomedical research, particularly concerning protein interaction networks and metabolic networks, the determination of whether links (interaction relationships) exist between nodes often relies on extensive experimental inference. Taking protein interaction networks as an example, about 80% of interaction relationships in yeast proteins remain undiscovered, and for humans, only 0.3% of known protein interactions have been identified. Due to the high experimental costs associated with uncovering these hidden links in such networks, the development of effective link prediction algorithms based on existing network structures and characteristics is of paramount importance. Utilizing these predictive outcomes to guide experiments can significantly increase the likelihood of successful experimental results, thereby reducing costs. Furthermore, it can accelerate the pace of revealing the intrinsic mechanisms within these networks, offering substantial academic and research value [5], [6], [7].
In the realms of theoretical research and modeling, link prediction serves not only as a crucial tool for studying network structures and their evolutionary patterns but also as an effective method for simulating complex systems and constructing models. For example, in the field of knowledge representation, knowledge graphs, as a form of complex network, possess remarkable expressive power and modeling flexibility. In these graphs, nodes represent entities or concepts from the real world, and each link (edge) corresponds to a piece of knowledge in reality, thus embodying a wealth of rules and logical meanings. This allows for the deduction of unexpressed knowledge in the knowledge graph based on predefined rules. For instance, knowing that ‘Tom is a cat’, we can derive numerous new pieces of knowledge using rules such as ‘cats have two ears, four legs, and come in various breeds’, without the need to explicitly detail each one in the knowledge graph. Therefore, knowledge graphs not only effectively model the real world but are also readily processed by computers, leading to their wide application in fields like question-answering systems and criminal investigations in public security. However, issues like incomplete data and information loss during the construction of knowledge graphs can lead to missing entities, attributes, and relationships, causing inaccuracies in knowledge inference. To address these issues, knowledge graph completion techniques have emerged. These techniques aim to predict unobserved relationships between entities in the knowledge graph and to forecast tail entity attributes based on head entity attributes. Fundamentally, this technique parallels link prediction in complex networks. Thus, link prediction contributes significantly to enhancing the accuracy of knowledge graph completion, enriching knowledge inference, and its applications [8], [9].
With the advancement of deep neural networks, GNNs have been widely applied to link prediction in complex networks. On one hand, GNN-based prediction methods, being node-centric, update and learn node representations through repeated exchanges of neighborhood information, and then utilize prediction components to learn the similarity of these representations for link prediction. This approach, however, does not fully leverage the end-to-end learning advantages of GNNs and cumulatively increases the overall model error, making it challenging to dynamically adjust the node representations provided to the prediction components for optimization during training. Moreover, link-centric prediction models lack effective methods in areas such as the extraction and isomorphism testing of local pattern closure subgraphs of links. This project aims to focus on links by concentrating on the extraction and isomorphism testing of link closure subgraphs, transforming link prediction into a classification of these subgraphs, and establishing a single-task optimization model. On the other hand, previous studies have shown that current GNNs exhibit a low-pass filtering effect when processing network data [10], often learning only the low-frequency information representing commonalities of nodes and neglecting the high-frequency information that reflects node differences. This limitation confines the application of existing GNNs primarily to link prediction in strongly assortative networks, while their predictive performance is restricted in highly disassortative networks [8], [9], [10], [11]. However, highly disassortative networks are prevalent in the real world and play a crucial role, such as in biological, technological, and financial networks. Link prediction in these networks not only addresses practical application issues but also allows for exploration of network formation and evolution at a micro-level.
To address the aforementioned challenges, this article adopts a link-centric approach, proposing a method for predicting links in complex networks by Integrating Enclosure Subgraphs with High-Frequency Graph Information (IESHGI). This approach indirectly facilitates link prediction through the classification of link enclosure subgraphs. The main contributions of this work are summarized as follows:
We extract the closure subgraph, consisting of two nodes associated with a link and their neighboring nodes that do not exceed
hops, for each link, representing the link by its closure subgraph, and transform the link prediction problem into a classification task of these subgraphs. This leads to the establishment of a single-task optimization model for link prediction.$k$ We construct a GNN to learn the representations of link closure subgraphs, which extracts not only the low-frequency information from node representations but also the high-frequency information that reflects node differences. Furthermore, by applying an attention mechanism, we integrate high and low-frequency graph information to realize a universal graph filter. This allows the model to effectively and adaptively aggregate the features of neighboring nodes.
Extensive experiments have been conducted on widely recognized benchmark datasets to validate the feasibility and effectiveness of the proposed link prediction method.
The remainder of this article is organized as follows. Section I introduces the research background, main challenges faced, and the primary work of this article related to link prediction. Section II presents an overview of the literature involved to the topic under consideration. Section III outlines the link prediction framework that incorporate enclosure subgraphs, high- and low-frequency graph information. Section IV describes the experimental settings, datasets, along with the presentation of experimental consequences and their analysis. Finally, we concludes the key findings and drawbacks of the proposed method, and shed light on the future research directions in the final section V.
Related Works
As an emerging technique in complex network analysis, the concept of GNN was first introduced by Gori et al. [12] and further elucidated by Scarselli et al. [7]. However, in a comprehensive synthesis of existing research on link prediction, Philip S. Yu et al. pointed out that while methods based on similarity for link prediction have been extensively studied, the application of GNNs in link prediction has received comparatively less attention [5]. This paper will analyze and summarize the current state of research in representation learning within networks and the construction of GNNs for link prediction.
A. Network Representation Learning
“Network representation learning is focused on embedding network nodes into a low-dimensional space, transforming high-dimensional sparse feature vectors into low-dimensional dense embedding vectors. Methods based on random walks in network representation learning generate contextual information for network nodes through these walks. Node sequences are then interpreted as sentences, and natural language processing techniques are applied for node embedding. Consequently, the more frequently two network nodes appear together in the same random walk, the more similar their embeddings become.
One of the most representative random walk-based network representation learning algorithms is DeepWalk, introduced by Perozzi et al. [13]. Its fundamental concept involves mapping the relationships and structural properties of nodes within a graph to a new vector space. In this space, nodes that are closer within the network are also closer in the vector space, thus transforming network data into vector space data through this optimization goal. Following this, feature vectors representing node structural information are concatenated with those representing node attribute information, and then used for downstream network data mining tasks, including link prediction [14]. However, LINE proposed by Tang et al. [15], seemingly does not utilize a random walk strategy. Nevertheless, literature [16] categorizes it under random walk approaches, primarily because LINE, similar to DeepWalk, employs a probabilistic loss function. This involves minimizing the empirical probability of nodes being connected and the distance in vectorized node similarity, considering both first and second-order similarities. This approach is inherently akin to the motivations of random walk strategies.
Furthermore, for strongly assortative network data, Xu et al. proposed the Graph Isomorphism Network (GIN) [17], which characterizes the discriminative ability of classical GNNs and their variants on assortative network data. GIN is proven not only to possess isomorphism testing capabilities as powerful as the Weisfeiler-Lehman test but also to perform exceptionally well on multiple network classification benchmark datasets. In terms of simultaneously learning network structure and embeddings, Chen et al. introduced an end-to-end learning framework, namely Deep Iterative and Adaptive Learning for Graph Neural Networks (DIAL-GNN) [18]. This framework converts the network structure learning problem into a similarity metric learning task, using an optimized regularization strategy to control the smoothness, connectivity, and sparsity of the generated network. Building on this foundation, they further proposed a new iterative method to search for hidden network structures, aiming to enhance the original network. This approach offers valuable guidance for constructing the link prediction Graph Neural Networks in this research topic.
In summary, network representation learning emphasizes the representation of network nodes and the preservation of network topology information in the embedding space, providing support for downstream link prediction tasks. However, current link prediction methods based on network representation learning rarely adopt a link-centric approach. They mainly focus on learning node representations in a node-centric manner and then utilize prediction components to calculate the similarity of these representations for link prediction. This approach not only accumulates the overall error of the model but also complicates the training process.
B. Construction of Gnns for Link Prediction
As a framework for deep learning on graphs, GNNs have been recognized for their potential in complex network link prediction. Yet, as noted by Philip Qiu et al., there is a relative scarcity of research in this area [11]. Baldassarre and Azizpour provided a general definition of Graph Networks (GNs) and focused their explanation on two main approaches: gradient-based and decomposition-based [20]. Their work, which concentrated on the interpretability of GNs, particularly for graph-based predictive tasks, offers valuable insights and inspiration for adapting link prediction in both assortative and disassortative complex networks, a focus of this study. To comprehensively learn node features in hierarchical graph-structured neural network models, information can be gleaned at various levels of the graph [21], [22]. The Capsule Graph Neural Network (CapsGNN) developed by Xinyi and Chen [22] stands out as one of the most representative models, employing the concept of capsules to address the limitations of existing graph embedding algorithms. By extracting node features in capsule form and utilizing a routing mechanism to gather vital information at different graph levels, CapsGNN generates multiple embeddings for each graph, capturing the macroscopic attributes of the entire graph in a data-driven manner from various perspectives. Our earlier work proposed a link-centric approach to link representation learning and GNN model training [19]. This model represents and learns link representations by ‘merging’ the representations of the two nodes associated with a link, defines an edge convolution layer, and constructs a link prediction GNN as shown in Figure 1 by stacking these layers. However, this model requires extracting node representations from the learned link representations and then computes link prediction based on the similarity of these node representations. Therefore, it accumulates errors in GNN training and link prediction, necessitating further refinement. However, combining the findings of Bo et al. published at AAAI 2021 [10], it is evident that current GNNs exhibit a low-pass filtering effect when processing network data, learning only low-frequency information and neglecting high-frequency information. This tendency limits GNNs to link prediction in strongly assortative networks and hinders their performance in highly disassortative networks. The smoothness of low-frequency information leads to GNN training retaining low-frequency features that reflect node commonalities, while high-frequency features that demonstrate node differences are overlooked. This paper will integrate the methods proposed by Bo et al. [10] and others to incorporate both high- and low-frequency graph information into the construction of graph link prediction neural networks.
Link Prediction Framework Integrating Enclosure Subgraph and High-Frequency Graph Information
This section will first introduce the symbols and task definitions related to link prediction. Subsequently, it will detail the scheme for extracting link enclosure subgraphs, the methods for high- and low-frequency graph information extraction, and the construction of Graph Neural Networks based on these elements. Finally, building on the aforementioned research, this section will outline a comprehensive framework for link prediction.
A. Preliminaries
1) Symbols
In our endeavor to enhance the clarity and depth of descriptions and explanations pertaining to the domain of link prediction, we have diligently compiled an extensive and detailed list of symbols, as shown in Table 1, which provide a clearer understanding of the complex concepts and methodologies employed in our research. Throughout this article, these symbols are used consistently, serving as a fundamental cornerstone for elucidating our theoretical and experimental approaches.
2) Link Prediction
Link prediction fundamentally aims to determine the existence or potential formation of links between two nodes. Within the framework of a specified graph
B. Link Enclosure Subgraphs Extraction
Guided by the SEAL proposed by Zhang and Chen [23], this article focuses on links within complex networks and extracts local pattern enclosure subgraphs of these links as the fundamental units for learning in a Graph Convolutional Neural Network (GCN) aimed at link prediction. We extract enclosure subgraphs not only for ‘observed’ links but also for ‘unobserved’ links, subsequently labeling the respective enclosure subgraphs’ categories as ‘1’ and ‘0’. This approach effectively translates the link prediction problem into a enclosure subgraph classification problem. To accommodate network types and node attributes, network embedding algorithms are utilized to generate initial representations of these link enclosure subgraphs. As illustrated in Figure 2, this process encompasses several key steps: the extraction of link enclosure subgraphs, the labeling of subgraph nodes, the dimension normalization of the subgraph adjacency matrix, and the generation of initial subgraph representations.
The process of extracting link closure subgraphs, labeling subgraph nodes, and generating initial subgraph representations constitutes a critical component of our methodology, where the edge ‘A—B’ is the ‘observed’ link and the ‘C
1) Enclosure Subgraphs Extraction Scheme
Motivated by the SEAL approach [23], this paper adopts a strategy of randomly sampling K-hop neighbors of link nodes to extract enclosure subgraphs. The process unfolds as follows:
Firstly, the method involves extracting enclosure subgraphs by randomly sampling K-hop (k=2 for simplicity in this article) neighbors of link-adjacent nodes. This process includes not only the extraction of closure subgraphs for ‘observed’ links (such as ‘A—B’ in Figure 2) but also for ‘unobserved’ links (like ‘C
Secondly, the nodes within the extracted enclosure subgraphs are labeled to achieve ‘sequential numbering’ of nodes. In networks, there often exist links with similar or identical roles, leading to isomorphic properties in their corresponding enclosure subgraphs. Building upon the Weisfeiler-Lehman algorithm [24], [25] and incorporating the randomness of enclosure subgraph extraction, this paper develops a node labeling algorithm with equivalent node labeling and isomorphism testing capabilities. The steps of this algorithm are illustrated in Figure 3. This approach ensures that enclosure subgraphs with isomorphic properties have adjacency matrices with similar node indices.
Node labeling process in enclosure subgraph of links, following the steps outlined in literatures [24] and [25]. Initially, all nodes are labeled with the same number, such as 1. Then, for each node, create a list of its neighboring nodes and represent it in the form of (node label, list of neighboring nodes label), such as (1, 11). Finally, based on the number of neighbors for each node, update its label to be the count of its neighbors; for example, the label of node (1, 11) is updated to 2. Repeat above process until each node within the enclosure subgraph has a unique label.
Finally, the dimensions of the enclosure subgraph adjacency matrices are ‘normalized’. Due to the sparsity, scale-free nature, and community structure characteristics of networks, there can be inconsistencies in the number of nodes in enclosure subgraphs formed by K-hop neighbors of link-adjacent nodes. This results in varying dimensions of the corresponding adjacency matrices for the enclosure subgraphs. In this paper, ‘n’ represents the final dimension of the subgraph adjacency matrix
2) Initial Representation Generation of Link Enclosure Subgraphs
To simultaneously capture the topological structure and attributes of enclosure subgraphs, and to learn and obtain the initial representation
C. Construction of GNN Integrating High- and Low-Frequency Graph Information
The adjacency matrix
Construction process of GCN layer for learning representations of link enclosure subgraphs, where
1) Extraction of High- and Low-Frequency Graph Information
For each link enclosure subgraph
Step One: Given that \begin{equation*} (\mathcal {F}\star \mathcal {X})_{\mathcal {G}}=\mathcal {U}((\mathcal {U}^{\top }\mathcal {F})\odot (\mathcal {U}^{\top }\mathcal {X}))=\mathcal {U}g_{\theta }\mathcal {U}^{\top }\mathcal {X} \tag {1}\end{equation*}
Herein,
Step Two: Inspired by Bo et al. [10], this paper employs high-pass and low-pass filters to extract high and low-frequency graph information from the features of nodes in enclosure subgraphs. The high-pass filter \begin{align*} {\mathcal {F}}_{hi}& =\alpha {\mathcal {I}}_{n}-\mathcal {D}^{-\frac {1}{2}}\mathcal {A}\mathcal {D}^{-\frac {1}{2}} \tag {2}\\ {\mathcal {F}}_{lo}& =\alpha {\mathcal {I}}_{n}+\mathcal {D}^{-\frac {1}{2}}\mathcal {A}\mathcal {D}^{-\frac {1}{2}} \tag {3}\end{align*}
\begin{align*} {\mathcal {X}}_{hi}& =({\mathcal {F}}_{hi}\star \mathcal {X})_{\mathcal {G}} = \mathcal {U}[(\alpha -1){\mathcal {I}}_{n}+\Lambda ]\mathcal {U}^{\top }\mathcal {X} \tag {4}\\ {\mathcal {X}}_{lo}& =({\mathcal {F}}_{lo}\star \mathcal {X})_{\mathcal {G}} = \mathcal {U}[(\alpha +1){\mathcal {I}}_{n}-\Lambda ]\mathcal {U}^{\top }\mathcal {X} \tag {5}\end{align*}
2) Universal Graph Filter Integrating High- and Low-Frequency Graph Information
Following the extraction of high- and low-frequency graph information from the link enclosure subgraph representation \begin{align*} w_{hi}^{i}& =softmax_{i}({\mathcal {X}}_{hi})=\frac {exp({\mathcal {X}}_{hi}^{i})}{\sum _{k\in \mathcal {N}(i)}exp({\mathcal {X}}_{hi}^{k})} \tag {6}\\ w_{lo}^{i}& =softmax_{i}({\mathcal {X}}_{lo})=\frac {exp({\mathcal {X}}_{lo}^{i})}{\sum _{k\in \mathcal {N}(i)}exp({\mathcal {X}}_{lo}^{k})} \tag {7}\end{align*}
It should be noted that \begin{equation*} \mathcal {H}^{(l)}= w_{hi}(({\mathcal {F}}_{hi}\star \mathcal {X})_{\mathcal {G}}) + w_{lo}(({\mathcal {F}}_{lo}\star \mathcal {X})_{\mathcal {G}}) +\mathcal {H}^{(l-1)} \tag {8}\end{equation*}
herein,
D. Link Prediction Framework Implemented through Link Enclosure Subgraph Classification
In this section, we delineate the construction of a comprehensive framework IESHGI for link prediction within complex networks. By stacking the aforementioned Graph Convolutional Network (GCN) layers and incorporating activation functions, pooling operations, and a classification function, we construct a Graph Convolutional Neural Network for the classification of link enclosure subgraphs, as depicted in Figure 5. This architecture enables the convolutional processing of the adjacency matrix and initial representations of link enclosure subgraphs, facilitating the acquisition of final subgraph representations. Consequently, this framework allows for the effective classification of link enclosure subgraphs, thereby enhancing our ability to predict links within complex networks.
The whole GCN designed for link prediction in complex networks and named IESHGI,combines both high- and low-frequency graph information, where
This framework not only leverages the structural and feature information inherent in the subgraphs but also optimizes the information flow through the network by applying non-linear transformations and reducing dimensionality where necessary. The use of pooling operations, in particular, aids in abstracting higher-level features from the convolutionally processed subgraphs, while the classification function translates these features into probabilistic predictions for subgraph categories. This methodology underscores the potential of deep learning techniques in unraveling the intricate patterns of connectivity that characterize complex networks, offering a robust tool for the prediction of new or missing links based on observable network dynamics.
Experiments and Discussion
To rigorously evaluate the IESHGI model and conduct a comparison with established baseline methods, we follow the experimental framework from our prior work [4]. Our method encompasses an in-depth analysis across various benchmark datasets to fully gauge the model’s efficacy. This section starts by detailing the benchmark datasets used, describes the baseline methods for comparison, and specifies the metrics for evaluating performance. Subsequently, we report our experimental findings and engage in a comparative discussion. Our aim is to showcase the IESHGI model’s robustness and dependability through this comprehensive examination.
A. Experimental Settings
The experimental framework for this study was meticulously crafted to guarantee the precision and reliability of our research work. We conducted our experiments on a Dell T640 workstation, a high-performance deep learning platform running the CentOS-7 operating system. This setup provides a stable and efficient platform for computational tasks. The workstation was equipped with a Tesla V100s GPU. For this research, CUDA 10.2 was utilized to leverage optimized support for graph deep learning frameworks and enhance GPU acceleration. The programming environment was unified under Python 3.7 to ensure seamless compatibility and facilitate the development of IESHGI. PyTorch version 1.11, known for its dynamic computational graph capability and comprehensive library support, was selected for IESHGI implementation. Additionally, we incorporated torch_geometric 2.1, a PyTorch-based library dedicated to graph deep learning. This library offers vital tools for the implementation and evaluation of the IESHGI model, bolstering our experimental setup with the necessary resources for cutting-edge research.
B. Dataset
Without loss of generality, while maintaining continuity with our previous research work [4], the dataset employed in this article remains consistent with the dataset utlized in our previous work [4], all sourced from the Open Graph Benchmark (OGB),1 as shown in Table 2, including ogbl-ppa [27], ogbl-collab [28] and ogbl-ddi [29]. We strictly followed the default partition settings provided by OGB for these datasets. This approach not only preserves the distinct size and characteristics of each dataset but also ensures consistency with the experimental methodologies outlined in [4], thereby maintaining the integrity and continuity of our research.
The ogbl-ppa is a type of undirected and unweighted graph. In this graph, nodes correspond to proteins originating from 58 distinct species. The edges within this graph signify biologically significant relationships between proteins, which can include physical interactions, co-expression patterns, homology, or genomic proximity [27].
The ogbl-collab is an undirected graph that captures a portion of the collaboration network among authors indexed by Microsoft Academic Graph. Each node in the graph represents an author, and the edges signify collaborations between these authors. All nodes in this dataset are associated with 128-dimensional features, which are derived by averaging the word embeddings of papers authored by these individuals. Additionally, each edge in the graph is accompanied by two pieces of meta-information: the year of collaboration and the edge weight, which reflects the number of co-authored papers published in that particular year. This graph can be understood as a dynamic multi-graph, allowing for the existence of multiple edges between two nodes if authors collaborate across multiple years [28].
The ogbl-ddi is a homogeneous, unweighted, undirected graph that depicts the drug-drug interaction network. In this graph, each node represents either an FDA-approved drug or an experimental drug. The edges in the graph signify interactions between these drugs, indicating scenarios where the combined effect of taking two drugs together significantly deviates from the expected effect of each drug acting independently. This network captures and visualizes the complex relationships and interactions between different drugs based on their observed joint effects [29].
C. Baseline Methods
In this study, we extend our previous research by further validating the effectiveness of our enhancements and comparing them with established baseline methods. To ensure a robust comparison, we employed the same baseline techniques as mentioned in literature [4], which include the classic Graph Convolutional Network(GCN) [30], Graph Attention Network(GAT) [31], GraphSAGE [32], EdgeConvNorm [19], and our own previously developed LVGANN [4]. It’s important to emphasize that our focus in using these GNNs is primarily on enclosure subgraph representation learning. For the link prediction task, which is indirectly approached through link enclosure subgraphs classification, we apply the log_softmax classifier to the pooled representations of these subgraphs, ensuring a consistent methodology for evaluating performance. The concise description of the baseline methods employed in this article are as follows.
GCN advances the application of convolutional operations beyond their traditional realm of regular grids, as seen in image processing, to accommodate the complexities of irregular graphs. Within this framework, graphs are processed by applying a feature transformation to each node, which incorporates the attributes of its neighbors. This pivotal process preserves the network’s local structure by aggregating neighboring features to forge a new representation for each node.
GAT introduces attention mechanisms to GNNs, revolutionizing how significance is allocated among nodes. By incorporating an attention mechanism, GAT independently assesses the importance of each neighboring node, acknowledging that different neighbors exert varying levels of influence. This innovation enables nodes to focus more on significant neighbors and less on those with minimal impact. As a result, GAT improves the model’s capacity to identify and learn from the most pertinent connections within the graph, facilitating more refined and effective node representations. This advancement underscores GAT’s role in enhancing the nuanced understanding and processing of graph data.
GraphSAGE an innovative adaptation of the GCN, addresses the critical issue of scalability. It creates node representations by selectively sampling and aggregating features, a departure from the GCN approach that requires processing all graph nodes in each forward pass. GraphSAGE stands out for its ability to efficiently train on large-scale graphs and produce embeddings for nodes not seen during training. This method greatly improves the model’s flexibility and scalability, establishing GraphSAGE as a formidable solution for navigating and analyzing vast and intricate graph networks.
EdgeConvNorm revolutionizes link representation learning by incorporating a specialized edge convolution operation, uniquely suited for distilling the essence of connections within a graph. This model further enhances the quality of link representations through a strategic normalization process, effectively addressing the prevalent challenge of over-smoothing often encountered in edge convolution-based link prediction models. A key feature of EdgeConvNorm is its deployment within a link prediction framework, employing multiple stacked edge convolutional layers. This structured approach allows the model to adeptly unravel and analyze intricate link characteristics, significantly boosting the precision and resilience of link prediction outcomes.
LVGANN introduces a nuanced analysis of link value grounded in network structure and presents a novel methodology for its estimation. This approach integrates link value into the design and training phases of a link prediction graph attention network, enhancing the precision of link predictions. Moreover, this integration offers a theoretical framework for interpreting the prediction outcomes, thereby enriching our understanding of link dynamics within networks. This advancement not only elevates the accuracy of link predictions but also contributes significantly to the theoretical underpinnings of network analysis.
D. Evaluation Metric
Evaluating the efficacy of link prediction models, \begin{equation*} Hits\text{@}n=\frac {1}{k}\sum p(e_{i}) \tag {9}\end{equation*}
The
E. Results Analysis and Discussion
For this study, the experimental settings of the baseline methods were carefully calibrated using the published source code from OGB. The model’s parameters were not further optimized, prioritizing the comparison of relative performances across different models. However, it is crucial to emphasize that this paper serves as an enhancement and refinement of our previous research [4]. To maintain continuity, comparability, and validate the effectiveness of the methodology introduced here, all baseline methods experimental consequences, apart from the IESHGI model’s experimental data, are sourced from our earlier studies [4]. This approach ensures a coherent and comparative framework for assessing the advancements presented in this work.
Besides, the final evaluation of performance is conducted by calculating the mean and standard deviation of the peak results from 10 iterations. Detailed findings, showcasing the comparative superiority of all baseline methods across diverse datasets, are systematically presented in Tables 3 to 5. This methodical approach underscores the robustness and consistency of our evaluation process, providing a clear demonstration of model efficacy.
In our analysis of the ogbl-collab dataset, we have adhered to the experimental protocol as described by OGB and broadened our methodology to include temporal aspects, utilizing the data splitting technique outlined in OGB. Our main objective is to forecast future author collaborations using historical data, placing a particular focus on prioritizing true collaborations.
Table 3 clearly demonstrates that the IESHGI method introduced in this paper outperforms other baseline methods, including our previously proposed LVGANN approach based on link value assessment. It achieves an average accuracy of 0.5356 and shows an average performance improvement of 6.3% compared to the LVGANN method. Given the dynamic and complex nature of the multi-graph dataset ogbl-collab, link enclosure subgraphs are more adept at capturing the local characteristics of links. They also delve into the high- and low-frequency information within the link enclosure subgraphs. This not only captures the similarity in subgraph node representations but also highlights the differences, effectively reflecting the intrinsic mechanisms through which nodes form links. These features enable IESHGI to more accurately predict the existence of a link between two nodes.
For the ogbl-ddi dataset, we employed a protein-target splitting method. This strategy aims to forecast drug-drug interactions based on data from established interactions previously. The experimental data in Table 4 clearly indicate that the IESHGI method introduced in this paper has achieved an average performance improvement of 17.7% over our previously proposed LVGANN method, attaining an average performance of 0.6780. Fortunately, compared to GraphSAGE, which also derives network representations by learning from the network’s local structure, our method has shown a notable performance enhancement. The ogbl-ddi dataset is notable for its dense network structure, featuring 4,267 nodes and an impressive 1,334,889 edges. However, GraphSAGE’s innovative node sampling strategy demonstrates significant effectiveness in addressing these challenges. By judiciously choosing nodes for the training process, GraphSAGE effectively navigates the complexities of the dataset’s large-scale graph. It addresses the computational intensity of updating gradients over the entire graph and enhances training efficiency, illustrating a strategic approach to handling high-density networks. Fortunately, the IESHGI method proposed in this paper also predicts links by indirectly learning from the local structural enclosure subgraphs of a network. Overall, for dense network, local structures are more prevalent and better reflect the network’s characteristics, making IESHGI particularly suited for these environments.
For the ogbl-ppa dataset, we adhere to the partitioning strategy outlined in the established OGB framework. The primary goal of this article is primarily aimed at predicting specific types of protein relationships, emphasizing the prediction of physical protein-protein interactions. These auxiliary connections have demonstrated a significant correlation with the targeted interactions, thereby boosting the reliability and accuracy of predictions within the realm of protein relationships. Experimental data from Table 5 reveal that, in comparison to the ogbl-collab and ogbl-ddi datasets, all methods, including IESHGI, did not achieve optimal results on ogbl-ppa, despite IESHGI showing a slight improvement over our previous method LVGANN. The primary reason lies in the high sparsity structure of ogbl-ppa, which owns 576,289 nodes and 30,326,273 edges, making local structures like enclosure subgraphs less prevalent than in ogbl-collab and ogbl-ddi. This suggests that existing methods may be more suited to dense networks and those with high assortative.
Synthesizing the experimental results and analyses, it is evident that the IESHGI method proposed in this paper demonstrates strong overall performance, outperforming our previously introduced LVGANN method as well as conventional baseline methods, thereby advancing and refining our earlier work [4]. Data from Tables 3 and 4 indicate that both the existing baseline methods and the newly introduced IESHGI are particularly effective in dense networks, where local structures like enclosure subgraphs are more prevalent and exhibit high clustering. Additionally, IESHGI excels in learning and updating network representations by not only capturing low-frequency information that reflects commonalities among nodes but also high-frequency information that highlights node differences. This comprehensive approach to learning network representations is one of the key reasons IESHGI surpasses our previous link prediction methodologies.
Additionally, based on the dataset statistics shown in Table 2, the datasets employed in this study are of moderate size and exhibits high sparsity. Firstly, a sparse network implies fewer connections between nodes, leading to longer and more dispersed paths for information propagation. In such complex networks, conventional GNNs may face challenges in effectively propagating and aggregating information, resulting in information loss or decay. Secondly, the node degree distribution in sparse networks may be more uneven, with a few highly connected central nodes and a majority of nodes with lower degrees, which can impact GNNs’ ability to recognize and learn the overall network structure and important nodes. Fortunately, the proposed IESHGI model in this article effectively integrates high-frequency and low-frequency information, improving the model’s generalization ability and applicability.
Conclusion
This study introduces an inovative framework for predicting links within complex networks, a vital endeavor for uncovering latent or forthcoming node relationships applicable across diverse fields. Our research proposes a unique strategy that conceptualizes link prediction as a classification task of enclosure subgraphs, encapsulating both observed and unobserved links. This strategy leverages both high- and low-frequency information within these subgraphs, integrated via an attention mechanism, to construct a GNN specifically designed for acquiring intricate subgraph representations. This innovative approach not only sidesteps direct link prediction but also substantially boosts the accuracy of classifying link enclosure subgraphs.
Our method stands out by rectifying the biases present in conventional GNNs, showcasing impressive performance enhancements and robust generalization across well-established benchmark datasets. The strategy of using enclosure subgraphs to represent links simplifies link prediction into a straightforward optimization challenge. It captures a broad range of frequency information, effectively amalgamates features of adjacent nodes with a versatile graph filter, and sets a strong foundation for future link prediction endeavors.
Despite these advances, there remain areas ripe for further exploration and enhancement, particularly concerning the time complexity, attention mechanism refinement, and adaptability to highly sparse complex networks. (1) Time Complexity: A key challenge with sophisticated GNN models, including ours, is their substantial computational demand, which escalates with the network’s size and intricacy. The incorporation of both high- and low-frequency data, along with an attention mechanism for detailed subgraph representation learning, significantly contributes to the model’s increased time complexity. Future efforts could aim at optimizing these computational aspects, perhaps through refined algorithmic methods or by harnessing advancements in parallel computing. Such improvements would aim to lower the overall time complexity, enhancing the model’s suitability for real-time analysis and application to extensive datasets. (2) Refinement of the Attention Mechanism: The utilization of an attention mechanism is central to our model’s enhanced performance, enabling more precise integration of features from enclosure subgraphs. However, the potential for optimizing this mechanism further remains vast. Future research might investigate more advanced or dynamic attention frameworks that adjust responsively to the network’s specific features or the task at hand. Enhancements in this area could yield more accurate link predictions by more effectively discerning the network’s structural nuances and node interrelations. (3) Adaptability to Sparse Complex Networks: Our model shows exceptional performance in dense networks, where enclosure subgraphs are common. Yet, its efficacy in highly sparse complex networks, which typify many real-world environments, could be improved. Future work should consider devising strategies or adaptations to the model that bolster its performance in sparse contexts. This may involve creative approaches to deduce or reconstruct local structures in such networks or introducing novel mechanisms to capture essential long-range dependencies more effectively.
In summary, this study propels the link prediction field and the use of GNNs for network analysis forward. Yet, the identified avenues for enhancement underscore the roadmap for future investigations. By tackling these identified challenges, forthcoming studies have the opportunity to expand GNNs’ utility further, affirming their role as comprehensive tools for navigating and understanding the complexities of vast and diverse networks.