
Implications of sparsity and high triangle density for graph representation learning
nel
logistic(hXi, Yji)
(the latter already in common use),
assigning two latent positions,
Xi
and
Yi
, to each node
i
. More recently, Stolman et al. (2022) have attained the
properties of interest using the softmax kernel over a low-
dimensional embedding.
It is important to appreciate that, in our approach, the in-
ner product is not an arbitrary kernel. It provides a default
choice, because under mild assumptions it can be used to
represent any other positive definite kernel, although typi-
cally in infinite dimension, by Mercer’s theorem, known in
machine-learning as the ‘kernel trick’ (Steinwart and Christ-
mann, 2008). We find a transparent relationship between
the eigenvalues of the kernel and the triangle density (see
the supplementary material — Lemma 7), which shows
negative eigenvalues can only reduce the expected number
of triangles in the graph. As a result, if a carefully crafted
kernel seems to reproduce high triangle density (Chanpuriya
et al., 2020), and negative eigenvalues aren’t the reason, the
likely problem with the RDPG is with its finite-dimension
assumption.
There would be little hope for practical inference in infinite
dimension without structure. Instead, this paper shows that
sparse graphs with high triangle density can be modelled
using an infinite-dimensional inner product model, with
representations on a low dimensional manifold.
An implication of this manifold structure, that we later ex-
plore in depth, is the plausibility of global versus local
embedding. Under levels of sparsity which are generally
deemed reasonable, the manifold does not get populated
everywhere, and perfect reconstruction is impossible. How-
ever, since a
d
-dimensional manifold (informally) locally
resembles
d
-dimensional Euclidean space, we can ‘zoom in’
to the subgraph containing nodes in a small neighbourhood
of the latent space, and form an embedding which represents
a flat approximation of the manifold over that region. This
encompasses, but generalises, the notion of finding local
community structure (Rohe and Qin, 2013; Spielman and
Teng, 2013; Chen et al., 2020). A key technical contribu-
tion of our work is to construct manifolds for which we can
simultaneously control: sparsity, triangle density, and the
on-manifold distribution. One point of the latter is that we
can place a uniform distribution on the manifold, and still
achieve sparse graphs with high triangle density. Because
there is no reasonable notion of community structure in this
case, we find evidence against the common perception that
triangles imply community structure.
Related work
Several earlier works have proposed infi-
nite dimensional versions of the RDPG and/or the hypothe-
sis that latent positions live on a low-dimensional manifold
(Tang et al., 2013; Athreya et al., 2021; Rubin-Delanchy,
2020; Lei, 2021; Whiteley et al., 2021). The main goal of
this paper, to generatively achieve sparsity and high triangle
density, was not considered in these papers and indeed they
contain several assumptions which make the goal impossi-
ble, including a finite rank or trace-class kernel. The paper
Whiteley et al. (2021) provides a crucial element of our
theory, allowing a manifold construction for which we can
control the on-manifold distribution. We principally focus
on triangle density, rather than (the formal definition of)
transitivity, to align our work with Seshadhri et al. (2020);
however, our graphs also achieve high transitivity.
2 ACHIEVING HIGH TRIANGLE
DENSITY IN SPARSE GRAPHS
Setup
We consider a sequence of simple undirected
graphs,
A(n)
, on nodes indexed
1, . . . , n
. We make no dis-
tinction between the graph and its adjacency matrix,
A(n)
,
which is binary,
n×n
, and symmetric, with the standard
convention that
A(n)
ij = 1
if nodes
i
and
j
have an edge, and
zero otherwise.
We will assign to each node,
i
, an infinite-dimensional latent
position,
X(n)
i∈RN
, from a probability distribution
Fn
whose support
Mn
(the precise definition of which comes
later) only allows inner products that are valid probabilities,
that is,
hx, yi:=
∞
X
k=1
xkyk∈[0,1] for all x, y ∈ Mn. (1)
For convenience, the superscript
(n)
will be suppressed
from the latent positions, but there is no correspondence
between the latent positions of graphs of different sizes, i.e.
Xi
associated with
A(n)
is not the same as
Xi
associated
with A(m), for n6=m.
Definition 1
(Infinite-dimensional random dot product
graph)
.
Let
X1, . . . , Xn
i.i.d.
∼Fn
. The matrix
A(n)
rep-
resents an infinite-dimensional random dot product graph
(IRDPG) if
A(n)ind
∼Bernoulli(hXi, Xji),
conditional on X1, . . . , Xn, for 1≤i≤j≤n.
This model has appeared, at different levels of generality,
in many previous works (Young and Scheinerman, 2007;
Tang et al., 2013; Athreya et al., 2017; Lei, 2021), and
encodes several assumptions which we review at the end
of this section. Of particular interest here is the trade-off
between two key properties: the graph sparsity factor,
ρn
,
and triangle density,
∆n
, where (He, 2015; Seshadhri et al.,
2020)
nρn=nZZ hx, yidFn(x)dFn(y);
n∆n=n
3ZZZhx, yihy, zihz, xidFn(x)dFn(y)dFn(z),