
Theoretical Guarantees for Permutation-Equivariant Quantum Neural Networks
Louis Schatzki,1, 2, ∗Martín Larocca,3, 4, †Quynh T. Nguyen,3, 5 Frédéric Sauvage,3and M. Cerezo1, ‡
1Information Sciences, Los Alamos National Laboratory, Los Alamos, New Mexico 87545, USA
2Electrical and Computer Engineering, University of Illinois Urbana-Champaign, Urbana, Illinois, 61801, USA
3Theoretical Division, Los Alamos National Laboratory, Los Alamos, New Mexico 87545, USA
4Center for Nonlinear Studies, Los Alamos National Laboratory, Los Alamos, New Mexico 87545, USA
5Harvard Quantum Initiative, Harvard University, Cambridge, Massachusetts 02138, USA
Despite the great promise of quantum machine learning models, there are several challenges
one must overcome before unlocking their full potential. For instance, models based on quantum
neural networks (QNNs) can suffer from excessive local minima and barren plateaus in their training
landscapes. Recently, the nascent field of geometric quantum machine learning (GQML) has emerged
as a potential solution to some of those issues. The key insight of GQML is that one should design
architectures, such as equivariant QNNs, encoding the symmetries of the problem at hand. Here,
we focus on problems with permutation symmetry (i.e., symmetry group Sn), and show how to
build Sn-equivariant QNNs. We provide an analytical study of their performance, proving that
they do not suffer from barren plateaus, quickly reach overparametrization, and generalize well from
small amounts of data. To verify our results, we perform numerical simulations for a graph state
classification task. Our work provides theoretical guarantees for equivariant QNNs, thus indicating
the power and potential of GQML.
INTRODUCTION
Symmetry studies and formalizes the invariance of ob-
jects under some set of operations. A wealth of theory has
gone into describing symmetries as mathematical entities
through the concept of groups and representations. While
the analysis of symmetries in nature has greatly improved
our understanding of the laws of physics, the study of sym-
metries in data has just recently gained momentum within
the framework of learning theory. In the past few years,
classical machine learning practitioners realized that mod-
els tend to perform better when constrained to respect the
underlying symmetries of the data. This has led to the
blossoming field of geometric deep learning [1–5], where
symmetries are incorporated as geometric priors into the
learning architectures, improving trainability and general-
ization performance [6–13].
The tremendous success of geometric deep learning has
recently inspired researchers to import these ideas to the
realm of quantum machine learning (QML) [14–16]. QML
is a new and exciting field at the intersection of classi-
cal machine learning, and quantum computing. By run-
ning routines in quantum hardware, and thus exploiting
the exponentially large dimension of the Hilbert space, the
hope is that QML algorithms can outperform their classical
counterparts when learning from data [17].
∗louisms2@illinois.edu
†The two first authors contributed equally.
‡cerezo@lanl.gov
The infusion of ideas from geometric deep learning
to QML has been termed "geometric quantum machine
learning"(GQML) [18–24]. GQML leverages the ma-
chinery of group and representation theory [25] to build
quantum architectures that encode symmetry information
about the problem at hand. For instance, when the
model is parametrized through a quantum neural network
(QNN) [16,26–28], GQML indicates that the layers of the
QNN should be equivariant under the action of the sym-
metry group associated to the dataset. That is, applying a
symmetry transformation on the input to the QNN layers
should be the same as applying it to its output.
One of the main goals of GQML is to create architec-
tures that solve, or at least significantly mitigate, some
of the known issues of standard symmetry non-preserving
QML models [16]. For instance, it has been shown that
the optimization landscapes of generic QNNs can exhibit
a large number of local minima [29–32], or be prone to
the barren plateau phenomenon [33–45] whereby the loss
function gradients vanish exponentially with the problem
size. Crucially, it is known that barren plateaus and
excessive local minima are connected to the expressibil-
ity [30,32,37,43,46] of the QNN, so that problem-agnostic
architectures are more likely to exhibit trainability issues.
In this sense, it is expected that following the GQML pro-
gram of baking symmetry directly into the algorithm, will
lead to models with sharp inductive biases that suitably
limit their expressibility and search space.
In this work we leverage the GQML toolbox to cre-
ate models that are permutation invariant, i.e., mod-
els whose outputs remain invariant under the action of
arXiv:2210.09974v3 [quant-ph] 14 Feb 2024