
PREVIEW 1
Towards Consistency and Complementarity: A
Multiview Graph Information Bottleneck Approach
Xiaolong Fan, Maoguo Gong, Senior Member, IEEE, Yue Wu, Member, IEEE,
Mingyang Zhang Member, IEEE, Hao Li, and Xiangming Jiang
Abstract—The empirical studies of Graph Neural Networks
(GNNs) broadly take the original node feature and adjacency
relationship as singleview input, ignoring the rich information
of multiple graph views. To circumvent this issue, the mul-
tiview graph analysis framework has been developed to fuse
graph information across views. How to model and integrate
shared (i.e. consistency) and view-specific (i.e. complementarity)
information is a key issue in multiview graph analysis. In
this paper, we propose a novel Multiview Variational Graph
Information Bottleneck (MVGIB) principle to maximize the
agreement for common representations and the disagreement for
view-specific representations. Under this principle, we formulate
the common and view-specific information bottleneck objectives
across multiviews by using constraints from mutual information.
However, these objectives are hard to directly optimize since
the mutual information is computationally intractable. To tackle
this challenge, we derive variational lower and upper bounds of
mutual information terms, and then instead optimize variational
bounds to find the approximate solutions for the information
objectives. Extensive experiments on graph benchmark datasets
demonstrate the superior effectiveness of the proposed method.
Index Terms—Graph data mining, multiview graph represen-
tation learning, graph neural networks, deep neural networks.
I. INTRODUCTION
AS a ubiquitous data structure, graph is capable of mod-
eling real-world systems in numerous domains, ranging
from social network analysis [1, 2], natural language process-
ing [3, 4], computer vision [5, 6], and other domains. In recent
years, as a powerful tool for learning and analyzing graph
data, Graph Neural Networks (GNNs) [7, 8, 9, 10, 11] have
received tremendous research attention and have been widely
employed for graph analysis tasks. The common graph neural
networks broadly follow the message passing framework [9]
which involves a message passing phase and a readout phase.
The message passing first iteratively aggregates the neighbor
representations of each node to generate new node represen-
tations, and then the readout phase capture the global graph
information from node representation space to generate the
graph representation. The success of GNNs can be attributed
to their ability to simultaneously exploit the rich information
X. Fan, M. Gong (corresponding author), M. Zhang, H. Li, and
X. Jiang are with the School of Electronic Engineering, Key Labo-
ratory of Intelligent Perception and Image Understanding, Ministry of
Education, Xidian University, Xi’an, Shaanxi province, China, 710071.
(e-mail: xiaolongfan@outlook.com; gong@ieee.org; haoli@xidian.edu.cn;
myzhang@xidian.edu.cn; xmjiang@xidian.edu.cn)
Y. Wu is with the School of Computer Science and Technol-
ogy, Xidian University, Xi’an, Shaanxi province, China, 710071. (e-mail:
ywu@xidian.edu.cn)
inherent in the singleview graph topology structure and input
node attributes.
However, for real-world applications, graph data are often
manifested as multiple types of sources or different topology
subsets, which can be naturally organized as multiview graph
data. For example, a multiplex network [12, 13] contains mul-
tiple systems of the same set of nodes, and there exists various
types of relationships among nodes, which can be seen as
different topology subsets. Furthermore, existing approaches
may confront challenges, such as difficulty in obtaining labeled
data and generalization bias under the supervised learning
setting [14]. One way to alleviate these issues is unsupervised
representation learning on graph. Therefore, how to integrate
different graph views into low-dimensional representations
for downstream tasks in an unsupervised manner becomes a
fundamental problem for graph representation learning.
Recently, several multiview graph representation learning
approaches have been proposed to effectively explore the
multiview graph data. For instance, Adaptive Multi-channel
Graph Convolutional Networks (AM-GCN) [15] show that
performing graph convolutional operation over both topology
and feature views can improve the performance of node repre-
sentation learning, where the feature view can be generated by
distance based K-Nearest Neighbor algorithm. Note that AM-
GCN works in the supervised learning manner and therefore
may suffer from the challenge of annotating graphs. Con-
trastive Multiview Graph Representation Learning (MVGRL)
[16] presents an unsupervised mutual information maximiza-
tion approach for learning node and graph representations by
contrasting graph views, where the additional graph view is
generated by structural augmentations such as Personalized
PageRank diffusion and heat diffusion strategies. Specifically,
MVGRL maximizes the mutual information between two
views by contrasting node representations from one view with
graph representation from the other view, but neglects to ex-
plicitly distinguish the common and view-specific information
across multiviews. Recent work [17] has shown that explicitly
modeling common and view-specific information can improve
the performance of multiview representation model.
In this paper, we focus on the in-depth analysis of multiview
graph representation learning and aim to answer the question
that how to extract common and view-specific information
in an unsupervised manner for graph representation learning.
Inspired by the idea of mutual information constraints in
information bottlenecks [18, 19, 20, 21], we first formulate
the information bottleneck objective to encourage that the
latent representation contains as much information as possible
arXiv:2210.05676v1 [cs.LG] 11 Oct 2022