Towards Prototype-Based Self-Explainable Graph Neural Network Enyan Dai

2025-05-06 0 0 963.13KB 9 页 10玖币
侵权投诉
Towards Prototype-Based Self-Explainable Graph Neural
Network
Enyan Dai
The Pennsylvania State University
emd5759@psu.edu
Suhang Wang
The Pennsylvania State University
szw494@psu.edu
ABSTRACT
Graph Neural Networks (GNNs) have shown great ability in mod-
eling graph-structured data for various domains. However, GNNs
are known as black-box models that lack interpretability. Without
understanding their inner working, we cannot fully trust them,
which largely limits their adoption in high-stake scenarios. Though
some initial eorts have been taken to interpret the predictions
of GNNs, they mainly focus on providing post-hoc explanations
using an additional explainer, which could misrepresent the true
inner working mechanism of the target GNN. The works on self-
explainable GNNs are rather limited. Therefore, we study a novel
problem of learning prototype-based self-explainable GNNs that
can simultaneously give accurate predictions and prototype-based
explanations on predictions. We design a framework which can
learn prototype graphs that capture representative patterns of each
class as class-level explanations. The learned prototypes are also
used to simultaneously make prediction for for a test instance and
provide instance-level explanation. Extensive experiments on real-
world and synthetic datasets show the eectiveness of the proposed
framework for both prediction accuracy and explanation quality.
ACM Reference Format:
Enyan Dai and Suhang Wang. 2022. Towards Prototype-Based Self-Explainable
Graph Neural Network. In Proceedings of ACM Conference (Conference’17).
ACM, New York, NY, USA, 9 pages. https://doi.org/10.1145/nnnnnnn.nnnnnnn
1 INTRODUCTION
Graph structured data such as trac networks, social networks,
and molecular graphs are very pervasive in real world. To model
the graph structured data for various applications such as drug
discovery [
15
], nancial analysis [
32
], and recommendation sys-
tem [
33
], various graph neural networks (GNNs) [
5
,
12
,
17
] have
been proposed and made remarkable achievements. The success
of GNNs relies on the message-passing mechanism, i.e., the node
representations in GNNs will aggregate the information from the
neighbors to capture the attribute and topology information. Many
message-passing mechanisms have been investigated to learn pow-
erful representations from graphs, facilitating various tasks such
as node classication [8, 17] and graph classication [35].
Permission to make digital or hard copies of all or part of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed
for prot or commercial advantage and that copies bear this notice and the full citation
on the rst page. Copyrights for components of this work owned by others than ACM
must be honored. Abstracting with credit is permitted. To copy otherwise, or republish,
to post on servers or to redistribute to lists, requires prior specic permission and/or a
fee. Request permissions from permissions@acm.org.
Conference’17, July 2017, Washington, DC, USA
©2022 Association for Computing Machinery.
ACM ISBN 978-x-xxxx-xxxx-x/YY/MM. . . $15.00
https://doi.org/10.1145/nnnnnnn.nnnnnnn
Figure 1: An illustration of self-explanation with prototypes
on classifying whether the test graph is cyclic.
Despite the great success of GNNs in modeling graphs, GNNs
have the same issue of lacking explainability as other deep learn-
ing models due to the high non-linearity in the model. In addi-
tion, the message-passing mechanism of GNNs that aggregates
neighborhood features to capture topology information makes it
more challenging to understand the predictions. The lacking of
explainability in GNNs will largely limit their adoption in critical
applications pertaining to fairness, privacy and safety. For instance,
a GNN model may be trained to explore the proprieties of various
drugs. However, due to the black-box characteristic of GNNs, it is
unknown whether the rules learned by GNN model is consistent
with the chemical rules in the real-world, which raises the concern
of applying predictions that may threaten the drug safety.
Extensive approaches [
24
,
30
] have been investigated to explain
trained neural networks or give self-explainable predictions on
independent and identically distributed (i.i.d) data such as images.
However, they fail to generalize to GNNs due to the utilization of
message-passing mechanism designed for relational information
preservation. Recently, some initial eorts [
14
,
22
,
38
,
41
,
43
] have
been taken to address the explainability issue of GNNs. For example,
GNNExplainer [
38
] explains the prediction of an instance by identi-
fying the crucial subgraph of the instance’s local graph. Model-level
explanation is also investigated by generating graph patterns that
maximize the prediction of each class by the target model [
41
].
However, most of existing GNN explainers focus on the post-hoc
explanations, i.e., learning an additional explainer to explain the
predictions of a trained GNN. Since the learned explainer cannot
have perfect delity to the original model, the post-hoc explana-
tions may misrepresent the true explanations of the GNNs [
28
].
Therefore, it is crucial to develop a self-explainable GNN, which
can simultaneously give predictions and explanations.
One promising direction of self-explainable GNN is to learn
prototype graphs of each class to present the key patterns of each
class and simultaneously conduct prediction and give explanations
with the learned prototypes. The prototypes can provide class-level
and instance-level self-explanations. Figure 1 gives an illustration
of the prototype-based self-explanations on a toy classication
problem. As shown in the gure, the problem is to predict whether
a graph is cyclic or not. Taking the class cyclic as an example, the
learned prototype graphs are typical patterns of cyclic graphs of
arXiv:2210.01974v1 [cs.LG] 5 Oct 2022
various sizes. Therefore, the learned prototype graphs of class cyclic
can provide class-level explanation to show representative graphs of
class cyclic. For a test graph
G𝑡
, we will match it with the prototype
graphs of each class to give the prediction. Specically, the instance-
level explanation for predicting
G𝑡
can be: “Graph
G𝑡
is classied
as cyclic, because it is most similar to the second prototype graph
of class cyclic." Though promising, learning representative graphs
for self-explainable classication remains an open problem.
Therefore, in this work,we investigate a novel problem of learn-
ing prototype-based self-explainable graph neural network. How-
ever, this is a non-trivial task. There are two main challenges: (
i
)
how to eciently learn high-quality prototypes that are representa-
tives of each class for class-level explanation. Though some existing
works [
6
,
21
] have studied prototype learning for self-explanations,
they are mainly proposed for i.i.d data. Recently, ProtGNN [
45
]
applies a Monte Carlo tree search to identify subgraphs from raw
graph as prototypes. However, the search algorithm is very time
consuming. And the prototypes are limited to the subgraphs in
the dataset, which might not be that representative; and (
ii
) how
to simultaneously give an accurate prediction and provide correct
prototype-based instance-level explanation. Dierent from images,
the matching process between the test graph and prototype graphs
cannot directly use simple metric such as Euclidean distance. More-
over, the supervision of the matching result is not available. How
to eectively leverage the classication supervision for prototype
learning and correct explanations needs further investigation.
In an attempt to address the above challenges, we develop a
novel
P
rototype-Based Self-E
x
plainable
GNN
(PxGNN)
1
. To e-
ciently obtain the prototype graphs, PxGNN adopts a prototype
graph generator to attain the prototype graphs from the learnable
prototype embeddings. A constraint on the learnable prototype
embeddings and self-supervision from graph reconstruction are
utilized to guarantee the quality of learned prototype embeddings
and generated prototype graphs, respectively. An encoder is de-
ployed to match the test graph with the generated prototype graphs
for self-explainable classication. Since representative prototype
graphs of a certain class is supposed to be similar to the test graphs
in the same class, the labels can provide implicit supervision to
ensure the representativeness of the prototype graphs and guide
the matching process. More specically, a novel classication loss is
proposed to simultaneously ensure the accuracy of prediction and
the quality of prototype-based instance-level explanation. And the
classication loss is utilized to jointly train the model and prototype
embeddings to learn prototypes well represent their corresponding
classes. In summary, our main contributions are:
We investigate a novel problem of learning prototype graphs for
self-explainable classication on graph-structured data;
We develop a new framework PxGNN, which learns an eective
prototype generator with self-supervision to obtain high-quality
prototype graphs for accurate predictions and explanations;
We construct a synthetic dataset which can quantitatively evalu-
ate the prototype-base explanation; and
Extensive experiments on both real-world and synthetic datasets
demonstrate the eectiveness of our PxGNN in learning repre-
sentative prototypes for accurate self-explainable classication.
1Code and datasets will be released upon acceptance
2 RELATED WORK
2.1 Graph Neural Networks
Graph Neural Networks (GNNs) [
4
,
17
,
31
,
37
] have shown great
ability for representation learning on graphs, which facilitate var-
ious applications such as trac analysis [
46
], recommendation
system [
37
], and drug generation [
4
]. Generally, existing GNNs [
7
9
,
12
,
17
,
19
,
31
,
36
] utilize a message-passing mechanism that a
node’s representation is updated by aggregating and combining the
features from its neighbors. For example, GCN [
17
] averages the
representations of neighbors and the target node followed by an
non-linear transformation. GAT [
31
] adopts an attention mecha-
nism to better aggregate the representations of the nodes from the
neighbors. Recently, various GNN models are proposed to further
improve the performance of GNNs [
7
,
8
,
16
,
20
,
26
,
39
,
47
]. For in-
stance, FastGCN [
7
] is proposed to alleviate the scalability issue of
GCN. In addition, some methods [
8
,
20
] focus on overcoming the
oversmoothing issue of GCN and design deep GNNs to incorporate
more hops of neighbors. Moreover, to facilitate the downstream
tasks that are short of labels, self-supervised GNNs [
16
,
26
,
39
,
47
]
are investigated to learn better representations.
2.2 Explainability of Graph Neural Networks
Despite the great success of graph neural networks, the problem
of lacking explainability hinders the adoption of GNNs to various
high-stake domains such as credit estimation. Though extensive
methods [
1
,
11
,
13
,
29
,
40
,
44
] have been proposed to explain neural
networks, they are overwhelmingly developed for i.i.d data such
as images and texts and cannot be directly applied to explain GNN
models. Recently, some works in explainability of GNNs are emerg-
ing [
3
,
10
,
22
,
25
,
38
,
41
]. The majority of these GNN explainers
give the explanations by extracting the crucial nodes, edges, and/or
node features. For instance, GNNExplainer [
38
] learns soft masks
for edges and node features to explain the predictions with the
identied subgraphs and features. PGExplainer [
22
] proposes to
combine the global view of GNNs to facilitate the extraction of im-
portant graphs by applying a parameterized explainer. XGNN [
41
]
generates representative graphs for a class as model-level explana-
tions for graph classication.
However, the aforementioned methods focus on post-hoc ex-
planations for a trained GNN, i.e., they usually require additional
explainer to explain the target GNN, which might misrepresent the
decision reasons of the model. There are very few initial eorts
for self-explainable GNNs [
10
,
45
], which aims to simultaneously
give predictions and explanations on the predictions. SE-GNN [
10
]
simultaneously give the predictions and explanations of a target
node by identifying the K-nearest labeled nodes. ProtGNN [
45
] is
the most similar work to ours, which nds subgraphs from the raw
graphs as prototypes to give self-explanations. However, ProtGNN
only focuses on graph classication and the computational cost
is very large due to the searching phase in nding the prototype
subgraphs. Our proposed method is inherently dierent from this
work: (
i
) we propose a novel prototype-based self-explainable GNN
that is eective in both node and graph-level classication tasks;
(
ii
) a prototype generator is deployed to eciently learn more
representative prototypes for self-explainable classication.
摘要:

TowardsPrototype-BasedSelf-ExplainableGraphNeuralNetworkEnyanDaiThePennsylvaniaStateUniversityemd5759@psu.eduSuhangWangThePennsylvaniaStateUniversityszw494@psu.eduABSTRACTGraphNeuralNetworks(GNNs)haveshowngreatabilityinmod-elinggraph-structureddataforvariousdomains.However,GNNsareknownasblack-boxmod...

展开>> 收起<<
Towards Prototype-Based Self-Explainable Graph Neural Network Enyan Dai.pdf

共9页,预览2页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:9 页 大小:963.13KB 格式:PDF 时间:2025-05-06

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 9
客服
关注