
2
representations and GNNs.
Graph pooling is a technique used to reduce the dimen-
sion and extract the features of graphs, which usually ap-
pear with the convolutional layers [48]. The most widely
used methods are graph clustering algorithms [49–52], as well
as some other ones which have been lately studied [53–56].
HaarPooling is a graph pooling operation to compress and
filter graph features [57], based on compressive Haar trans-
forms. One of its important characteristics is, the basis for
forming a Haar matrix is computed by a clustering step from
the input graph, which means additional input-related infor-
mation can be passed to the ML process via the Haar matrix.
For quark-gluon tagging using GNNs, HaarPooling makes it
possible to embed extra particle features to filter and enhance
the message passing.
In our work, we combine HaarPooling with MPNN to
build a new network structure, called HaarPooling Message
Passing neural network (HMPNet). On one hand, jet events
are transformed into a graph representation as input for GNN,
and the tagging can be achieved by training with the pro-
cess of message passing and self-updating [37,39]. On the
other hand, in the updating process of the algorithm, the ad-
ditional particle feature is embedded through the compres-
sive Haar basis matrix of pooling, which makes the extraction
and classification of features more relevant to the input. This
means the pooling for compression also becomes an opera-
tion for adding fine information of input. For test, we imple-
ment the HMPNet to the quark-gluon tagging of the process
pp →Z/γ∗+j+X→µ+µ−+j+X, and use different
particle features such as absolute energy log E, transverse mo-
mentum log pT, the relative coordinates (∆η, ∆ϕ), the mixed
ones (log E, log pT)and (log E, log pT,∆η, ∆ϕ)to generate
the Haar matrix by clustering the input, respectively. We anal-
yse the influences of different particle features, and compare
the results of log pTwith the counterparts of other algorithms,
which shows a remarkable improvement of performance.
The main structure of this paper is as follows. In Section
II.1, the graph representation of jets will be given. Section II.2
gives the method of MPNN. Section II.3 includes the concep-
tions of graph pooling and Haar matrix. In Section II.4, the
method of embedding particle features to Haar matrix is il-
lustrated. In Section II.5, we explain the detailed process of
HMPNet. In Section III.1, the input data and settings of HMP-
Net are listed. Section III.2 shows our major findings. Section
IV is the conclusion of this work.
II. METHODOLOGY
II.1. Graph representation of jets
In the language of GNNs, an undirected graph G=
{V,E,X,W} is defined with nodes (vertices) V, edges E,
weights of nodes Xand of edges W. Each node vi∈ V has
its feature vector xi∈ X, and for the edge weight W, it is
always given in the form of an weight matrix dij in which the
element is given for the edge between i-th and j-th nodes in
the graph. And the number of nodes is defined as N=|V|.
Usually, the information of a jet reconstructed from de-
tectors in high-energy collision includes: the three Cartesian
coordinates of the momentum (px, py, pz), the absolute en-
ergy E, the pseudorapidity η, the azimuthal angle ϕ, the trans-
verse momentum pTand so forth. For the feature vectors xi,
we use 10 variables of jet information as components of xi
similar to Ref. [39], as shown in Table. I. The dynamic infor-
mation of objects includes log pT,log E, the relative energy
log E
E(jet)and the relative transverse log pT
pT(jet). In addition,
qdenotes the electric charge of object and the rest four fea-
tures are particles identity (PID) information. The dimension
of xiis N×dx, where dx= 10 is the dimension of the feature
space.
For graph representation, we also need to identify a pa-
rameter as the edge weight dij . From the point view of jet
axis, the relative distance ∆R=q∆η2
ij + ∆ϕ2
ij from the
jet center is a suitable choice, where the relative coordinates
∆ηij =ηi−ηjand ∆ϕij =ϕi−ϕjdenote the angle dif-
ference between the i-th with j-th particle in jet axis. By the
definition of ∆R, the edge weight is given by,
dij =q∆η2
ij + ∆ϕ2
ij .(1)
As an illustration, we show the graph events of the process
pp →Z/γ∗+j+X→µ+µ−+j+Xby Monte Carlo sim-
ulations in Fig. 1. As a graph representation with Nnodes,
each component of xiis a vector of dxelements of jet infor-
mation, with N= 9 and dx= 10. So dij is an N×N-
dimensional symmetric, matrix with all the diagonal elements
being 0. Since ϕis not encoded in the node features, the graph
representation is invariant under rotation in ϕ.
II.2. MPNN algorithm
The flexible and complete feature of graph makes it a nat-
ural and promising representation of jets; on the other hand,
to choose a paired algorithm of GNNs also requires careful
thought. Message Passing Neural Networks(MPNN) is in-
troduced as a powerful and efficient supervised algorithm of
GNNs which can learn geometric representations as well, es-
pecially the edge features dij [38,58]. By finding the opti-
mized parameters in the nonlinear network model via training,
one can obtain the classification as output of MPNN, from the
input graph representation of jets.
To start the process of MPNN, the feature vectors xi∈
RN×dxare embedded into a matrix consisting of higher di-
mensional state vectors s(0)
i∈RN×dswith ds> dx, by an
embedding function fe:
s(0)
i=fe(xi).(2)
Here s(0)
iis only related to xiwithout any information of the
graph structure. To encode the whole event graph into each
node state vector, message vector m(t)
iis introduced to pass
the message of s(t−1)
iand edge weight dij via the message