
INGNN: Uplifting Message Passing Neural Network with Graph Original Information
demonstrate that, simply exploiting the graph original in-
formation rather than deliberately handcrafting features can
improve the performance of MPNNs while preserving their
linear scalability. The two types of graph original informa-
tion, graph structures and node features, can be used both
together and separately, leading to three features including
ego-node features,graph structure features, and aggregated
neighborhood features. MPNNs, however, leverage only the
last feature for downstream tasks. We find that all three fea-
tures are necessary to improve the expressive power or the
generalizability of MPNNs and prove it both theoretically
and empirically. Based on the insight, we further propose an
INformation-enhanced MPNN (INGNN), which adaptively
fuses these three features to derive nodes’ representation and
achieves outstanding accuracy compared to state-of-the-art
baselines. We summarize our main contributions as follows:
•Theoretical Findings.
We show theoretically that the
expressiveness of an MPNN that integrates graph struc-
ture features is strictly better than
1&2
-WL and no
less powerful than
3
-WL. We also prove that the node
misclassification rate for graphs in some cases can be
depressed by utilizing ego-node features separately
from MPNNs. We empirically verify the necessity of
the three features and show that the node classifica-
tion accuracy of start-of-the-art models, GCN (Kipf &
Welling,2016), H2GCN (Zhu et al.,2020), and LINKX
(Lim et al.,2021a) can be improved by adding some
of the three identified features that they do not have
originally.
•INGNN.
We propose INGNN, a powerful GNN which
uplifts MPNNs’ expressiveness and accuracy by inte-
grating graph structure features and ego-node features.
The features are fused automatically with adaptive fea-
ture fusion under a bi-level optimization framework.
Our ablation study demonstrates that all three features,
the fusing mechanism, and the optimization procedure
play indispensable roles on different types of graphs.
•Evaluation.
We conduct extensive experiments to
compare INGNN with state-of-the-art GNNs using
synthetic and real graph benchmarks that cover the full
homophily spectrum. INGNN outperforms 12 baseline
models and achieves an average rank of 1.78 on 9 real
datasets. INGNN achieves higher node classification
accuracy,
4.17%
higher than GCN (Kipf & Welling,
2016),
4.01%
higher than MIXHOP (Abu-El-Haija
et al.,2019), and
3.43%
higher than GPR-GNN (Chien
et al.,2020), on average over the real datasets.
2. Related Works
Regular MPNNs.
Many GNNs (Niepert et al.,2016;
Hamilton et al.,2017;Monti et al.,2017;Veli
ˇ
ckovi
´
c et al.,
2017;Gao et al.,2018;Xu et al.,2018;Wang et al.,2019)
fall into the message passing framework (Gilmer et al.,2017;
Battaglia et al.,2018), which iteratively transforms and prop-
agate the messages from the spatial neighborhoods through
the graph topology to update the embedding of a target node.
To name a few, GCN (Kipf & Welling,2016) designs a layer-
wise propagation rule based on a first-order approximation
of spectral convolutions on graphs. GraphSAGE (Hamilton
et al.,2017) extends GCN by introducing a recursive node-
wise sampling scheme to improve the scalability. Graph
attention networks (GAT) (Veli
ˇ
ckovi
´
c et al.,2017) enhances
GCN with the attention mechanism (Vaswani et al.,2017).
Improve Expressiveness.
Several works attempt to im-
prove the expressiveness of GNNs. The first line of research
mimics the high-order WL tests such as 1-2-3 GNN (Morris
et al.,2019), PPGN (Maron et al.,2019), ring-GNN (Chen
et al.,2019). They usually require exponentially increas-
ing space and time complexity w.r.t. the number of nodes
in a graph and thus cannot be generalized to large-scale
graphs. The second line of research maintains the linear
scalability of more powerful GNNs by adding features, such
as subgraph counts (Bouritsas et al.,2022), distance encod-
ing (Li et al.,2020), and random features (Abboud et al.,
2020;Sato et al.,2021). The handcrafted features may intro-
duce inductive bias that hurts generalization and also lose
the premise of automatic learning (Zhao et al.,2021). The
third line of work tries to design more expressive GNNs
by involving high-order neighbors. For example, H2GCN
(Zhu et al.,2020) identifies ego- and neighbor-embedding
separation, second-order neighbors, and the combination of
intermediate representations to boost representation learn-
ing. MixHop (Abu-El-Haija et al.,2019) proposes a graph
convolutional layer that utilizes multiple powers of the adja-
cency matrix to learn general mixed neighborhood informa-
tion. Similarly, Generalized PageRank (GPR-) GNN (Chien
et al.,2020) adaptively controls the contribution of different
propagation steps. In this paper, we propose to improve
MPNNs’ expressiveness by integrating the original graph
structure features so that we can retain the linear complexity
and the automatic learning property of MPNNs.
Improve Generalizability.
Recent works start to pay at-
tention to improving the GNNs’ generalizability, including
overcoming over-smoothing and handling graphs with var-
ious homophily. DAGNN (Liu et al.,2020) proposes to
decouple the transformation and the propagation operation
to increase receptive fields. APPNP (Klicpera et al.,2018)
leverages personalized PageRank to improve the propaga-
tion scheme. Later, BernNet (He et al.,2021) learns an
arbitrary spectral filter such as the ones in GCN, DAGNN,
and APPNP via Bernstein polynomial approximation. Sim-
ilarly, AdaGNN (Dong et al.,2021) introduces trainable
filters to capture the varying importance of different fre-
quency components. Considering graph heterophily, ACM-
GCN (Luan et al.,2022) proposes a multi-channel mixing