2 Related Work
MPNN limitations. Given the understanding that MPNNs have expressiveness upper-bounded by
1-WL [61], several researchers investigated what else MPNNs cannot learn. To this end, Chen et al.
[15] showed that MPNNs cannot count induced connected substructures of 3 or more nodes, while
along with [15], Arvind et al. [3] showed that MPNNs can only count star-shaped patterns. Loukas
[38] further proved several results regarding decision problems on graphs (e.g. subgraph verification,
cycle detection), finding that MPNNs cannot solve these problems unless strict conditions of depth
and width are met. Moreover, Garg et al. [22] showed that many standard MPNNs cannot compute
properties such as longest or shortest cycles, diameters or existence of cliques.
Improving expressivity. Several works aim to improve expressiveness limitations of MPNNs. One
approach is to inject features into the MPNN aggregation, motivated by Loukas [38] who showed
that MPNNs can be universal approximators when nodes are sufficiently distinguishable. Sato et al.
[53] show that injecting random features can better enable MPNNs to solve problems like minimum
dominating set and maximum matching. You et al. [63] inject cycle counts as node features, moti-
vated by the limitation of MPNNs not being able to count cycles. Others proposed utilizing subgraph
isomorphism counting to empower MPNNs with substructure information they cannot learn [9,10].
Earlier work by You et al. [62] adds positional features to distinguish node which naïve MPNN
embeddings would not. Recently, several works also propose utilizing subgraph aggregation; Zhang
et al. [65], Zhao et al. [68] and Bevilacqua et al. [8] propose subgraph variants of the WL test which
are no less powerful than 3-WL. Recently a following up work [21] shows that rooted subgraph
based extension with 1-WL as kernel is bounded by 3-WL. The community is yet exploring several
distinct avenues in overcoming limitations: Murphy et al. [46] propose a relational pooling mecha-
nism which sums over all permutations of a permutation-sensitive function to achieve above-1-WL
expressiveness. Balcilar et al. [6] generalizes spatial and spectral MPNNs and shows that instances
of spectral MPNNs are more powerful than 1-WL. Azizian et al. [4] and Geerts et al. [24] unify
expressivity and approximation ability results for existing GNNs.
k-WL-inspired GNNs. The k-WL test captures higher-order interactions in graph data by consider-
ing all k-tuples, i.e., size kordered multisets defined over the set of nodes. While highly expressive,
it does not scale to practical graphs beyond a very small k(k= 3 pushes the practical limit even
for small graphs). However, several works propose designs which can achieve k-WL in theory and
strive to make them practical. Maron et al. [41] proposes a general class of invariant graph networks
k-IGNs having exact k-WL expressivity [23], while being not scalable. Morris et al. have several
work on k-WL and its variants like k-LWL [42], k-GNN [43], and δ-k-WL-GNN [44]. Both k-LWL
and k-GNN use a variant of k-WL that only considers k-sets and are strictly weaker than k-WL but
much more practical. In our paper we claim that k(≤)-sets should be used with additional designs
to keep the best expressivity while remaining practical. The δ-k-WL-GNN works with k-tuples but
sparsifies the connections among tuples by considering locality. Qian et al. [48] extends the k-tuple
to subgraph network but has same scalability bottleneck as k-WL. A recent concurrent work Spe-
qNet [45] proposes to reduce number of tuples by restricting the number of connected components
of tuples’ induced subgraphs. The idea is independently explored in our paper in Sec.3.4. All four
variants proposed by Morris et al. do not have realizations beyond k > 3in their experiments while
we manage to reach k= 10. Interestingly, a recent work [34] links graph transformer with k-IGN
that having better (or equal) expressivity, however is still not scalable.
3 A practical progressively-expressive isomorphism test: (k, c)(≤)-SETWL
We first motivate our method (k, c)(≤)-SETGNN from the GI testing perspective by introducing
(k, c)(≤)-SETWL. We first introduce notation and background of k-WL (Sec. 3.1). Next, we show
how to increase the practicality of k-WL without reducing much of its expressivity, by removing
node ordering (Sec. 3.2), node repetitions (Sec. 3.3), and leveraging graph sparsity (Sec. 3.4). We
then give the complexity analysis (Sec. 3.5). We close the section with extending the idea to k-FWL
which is as expressive as (k+1)-WL (Sec. 3.6). All proofs are included in Appendix A.4.
Notation: Let G= (V(G), E(G), lG)be an undirected, colored graph with nodes V(G), edges
E(G), and a color-labeling function lG:V(G)→Cwhere C={c1, ..., cd}denotes a set of
ddistinct colors. Let [n] = {1,2,3, ..., n}. Let (·)denote a tuple (ordered multiset), {{·}} de-
note a multiset (set which allows repetition), and {·} denote a set. We define −→
v= (v1, ..., vk)
as a k-tuple, ˜
v={{v1, ..., vk}} as a k-multiset, and ˆ
v={v1, ..., vk}as a k-set. Let −→
v[x/i] =
3