Line Graph Contrastive Learning for Link Prediction

2025-05-03 0 0 1.44MB 37 页 10玖币
侵权投诉
Highlights
Line Graph Contrastive Learning for Link Prediction
Zehua Zhang, Shilin Sun, Guixiang Ma, Caiming Zhong
We design a novel contrastive learning framework based on line graph
to be suitable for link prediction on sparse and dense graphs.
We propose a cross-scale contrastive learning strategy to maximize the
mutual information between subgraph and line graph.
The dual perspectives contrastive progress to some extent avoids the
problem of inconsistent prediction on the similarity based methods with
single view.
Our comprehensive experiments on six datasets from diverse areas
demonstrate that our model has better performance on generalization
and robustness than the SOTA methods.
arXiv:2210.13795v2 [cs.LG] 8 Mar 2023
Line Graph Contrastive Learning for Link Prediction
Zehua Zhanga,, Shilin Suna,∗∗, Guixiang Mab, Caiming Zhongc
aCollege of Information and Computer,Taiyuan University of Technology,Yuci District,
Jinzhong, 030600, Shanxi, China
bIntel Labs, Hillsboro, Oregon, 97229,USA
cCollege of Science and Technology, Ning bo University,Cixi, Ningbo, 315300, Zhejiang,
China
Abstract
Link prediction tasks focus on predicting possible future connections. Most
existing researches measure the likelihood of links by different similarity
scores on node pairs and predict links between nodes. However, the similarity-
based approaches have some challenges in information loss on nodes and
generalization ability on similarity indexes. To address the above issues, we
propose a Line Graph Contrastive Learning(LGCL) method to obtain rich
information with multiple perspectives. LGCL obtains a subgraph view by
h-hop subgraph sampling with target node pairs. After transforming the
sampled subgraph into a line graph, the link prediction task is converted
into a node classification task, which graph convolution progress can learn
edge embeddings from graphs more effectively. Then we design a novel cross-
scale contrastive learning framework on the line graph and the subgraph to
maximize the mutual information of them, so that fuses the structure and
Co-first authors and Corresponding author
∗∗ Co-first authors
Email addresses: zhangzehua@tyut.edu.cn (Zehua Zhang), shilin_sun01@163.com
(Shilin Sun), guixiang.ma@intel.com (Guixiang Ma), zhongcaiming@nbu.edu.cn
(Caiming Zhong)
Preprint submitted to Pattern Recognition March 9, 2023
feature information. The experimental results demonstrate that the proposed
LGCL outperforms the state-of-the-art methods and has better performance
on generalization and robustness.
Keywords:
Line Graph, Contrastive Learning, Link Prediction, Node Classification,
Mutual Information
1. Introduction
Link prediction task is based on the topological definition of the network
to predict the existence of links between nodes. It has been applied to various
fields, such as product recommendations [1], biological molecule interaction
prediction [2], traffic forecasting [3], etc.
The current research on link prediction usually follows a kind of human
intuition that the more similar the attributes or topological structure of two
nodes are, the more likely that they have interactions with each other. Based
on the common similarity principle, several network similarity methods have
been proposed for link prediction task [4]. And they are designed by minimiz-
ing the pointwise mutual information (PMI) of co-occurring nodes in random
walk [5]. Besides, the idea of multi-level analysis is introduced to deal with
graph structure data from the local and global levels. TOME [6] proposes
two refinement processes to obtain local information of cluster structure and
incorporates some global information into the matrix with path-based trans-
formation. Thereby, network similarity methods can also be classified from
multiple perspectives of the local and the global. The Node Clustering Co-
efficient [7] evaluates the clustering coefficients of all common neighbors of
2
the target node pair and sums them to obtain the final similarity score of
the node pair. Such methods can effectively handle link prediction task in
dynamic networks, such as traffic networks. Whereas, extracting only local
information will limit the ability to capture global similarities between nodes.
In contrast, other studies utilize global topological information of a network
to score the similarity of nodes, such as Katz [8], Random Walk with Restart
(RWR) [9], and Rooted Pagerank [10]. Except on sparsely unbalanced net-
works, the global structure based methods have shown better performance
than based on local information. Furthermore, the global methods are not
suitable for large-scale networks, specially with dense connections, due to
huge computational costs. In addition, the major challenge on these simi-
larity based methods does not avoid similarity measurement selection and
generalization limitation of a single index.
Obviously, another idea to improve the prediction accuracy can be de-
rived from edge information on graph to mine deeply the data to obtain
richer information. With the development of deep learning on graph data,
researchers pay more attention to graph representation learning methods with
the ability of learning graph topological information and enhancing node fea-
tures [11]. Specially, the feature learning process for the node is based on
the assumption that nodes with similar embedding representations will dis-
play similar structures. For example, HOGCN [2] adopts different distance
features on neighbors and shows excellent robustness in sparse interaction
networks. Wang et al. [12] introduce the HAS method via Heterogeneous
graph data Augmentation and node Similarity to solve sparse imbalanced
link prediction. What’s more, it is noteworthy that the representation and
3
structure dual similarity assumption is not universal. For instance, some
proteins with similar characteristics but may have a lower probability of con-
nections [13]. To address such issues, SEAL [14] converts the link prediction
task into a graph classification task by extracting the subgraphs around the
target links. Only the node information pooling is used to predict links in
SEAL, the loss of node information will bring disturbance to the accuracy
on prediction. Therefore, reducing the information loss has become another
challenge for current graph neural network-based methods.
In comparison, LGLP [15] transforms prediction task into a node classi-
fication task, by combining line graphs with graph neural networks for link
prediction to improve information transfer efficiency. LGLP can alleviate
the problem of inefficient learning with sparse data, whereas the added edges
during information transfer may also bring in noise. So it will limit the
model’s performance to ignore the balance of different levels of information.
Everything has two sides and the line graph transform is not an exception.
On dense graphs, the noise generated by the excessive edges of line graph
can adversely impact the prediction results [16]. By contrast, the informa-
tion generated by the increased edges of the transformation can improve the
prediction accuracy. So our motivation comes from how to make up for the
lack of line graph conversion and improve the performance by self-supervised
learning without additional information.
It is hard or expensive to acquire data labels in many practical appli-
cations, whereas contrastive learning is an excellent self-supervised learning
method to improve model performance with less labels. Current studies
focus on designing diverse graph augmentation strategies to yield various
4
摘要:

HighlightsLineGraphContrastiveLearningforLinkPredictionZehuaZhang,ShilinSun,GuixiangMa,CaimingZhongˆWedesignanovelcontrastivelearningframeworkbasedonlinegraphtobesuitableforlinkpredictiononsparseanddensegraphs.ˆWeproposeacross-scalecontrastivelearningstrategytomaximizethemutualinformationbetweensubg...

展开>> 收起<<
Line Graph Contrastive Learning for Link Prediction.pdf

共37页,预览5页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:37 页 大小:1.44MB 格式:PDF 时间:2025-05-03

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 37
客服
关注