
Stochastic noise can be helpful for variational quantum algorithms
Junyu Liu,
1, 2, 3, 4
Frederik Wilde,
5
Antonio Anna Mele,
5
Liang Jiang,
1, 2
and Jens Eisert
5, 6, 7
1
Pritzker School of Molecular Engineering, The University of Chicago, Chicago, IL 60637, USA
2
Chicago Quantum Exchange, Chicago, IL 60637, USA
3
Kadanoff Center for Theoretical Physics, The University of Chicago, Chicago, IL 60637, USA
4
qBraid Co., Chicago, IL 60615, USA
5
Dahlem Center for Complex Quantum Systems, Freie Universität Berlin, Berlin, 14195, Germany
6
Helmholtz-Zentrum Berlin für Materialien und Energie, 14109 Berlin, Germany
7
Fraunhofer Heinrich Hertz Institute, 10587 Berlin, Germany
Saddle points constitute a crucial challenge for first-
order gradient descent algorithms. In notions of classical
machine learning, they are avoided for example by means
of stochastic gradient descent methods. In this work, we
provide evidence that the saddle points problem can be
naturally avoided in variational quantum algorithms by ex-
ploiting the presence of stochasticity. We prove convergence
guarantees and present practical examples in numerical
simulations and on quantum hardware. We argue that
the natural stochasticity of variational algorithms can be
beneficial for avoiding strict saddle points, i.e., those saddle
points with at least one negative Hessian eigenvalue. This
insight that some levels of shot noise could help is expected
to add a new perspective to notions of near-term variational
quantum algorithms.
Quantum computing has for many years been a hugely inspir-
ing theoretical idea. Already in the 1980ies it was suggested
that quantum devices could possibly have superior computa-
tional capabilities over computers operating based on classical
laws [
1
,
2
]. It is a relatively recent development that devices
have been devised that may indeed have computational capa-
bilities beyond classical means [
3
–
6
]. These devices are going
substantially beyond what was possible not long ago. And still,
they are unavoidably noisy and imperfect for many years to
come. The quantum devices that are available today and pre-
sumably will be in the near future are often conceived as hybrid
quantum devices running variational quantum algorithms [
7
],
where a quantum circuit is addressed by a substantially larger
surrounding classical circuit. This classical circuit takes mea-
surements from the quantum device and appropriately varies
variational parameters of the quantum device in an update.
Large classes of variational quantum eigensolvers (VQE), the
quantum approximate optimization algorithm (QAOA) and
models for quantum-assisted machine learning are thought to
operate along those lines, based on suitable loss functions to
be minimized [
8
–
14
]. In fact, many near-term quantum algo-
rithms in the era of noisy intermediate-scale quantum (NISQ)
computing [
15
] belong to the class of variational quantum al-
gorithms. While this is an exciting development, it puts a lot of
burden to understanding how reasonable and practical classical
control can be conceived.
Generally, when the optimization space is high-dimensional
updates of the variational parameters are done via gradient eval-
uations [
16
–
19
], while zeroth-order and second-order methods
are, in principle, also applicable, but typically only up to a
limited number of parameters. This makes a lot of sense, as
one may think that going downhill in a variational quantum
algorithms is a good idea. That said, the concomitant classical
optimization problems are generally not convex optimization
problems and the variational landscapes are marred by glob-
ally suboptimal local optima and saddle points. In fact, it is
known that the problems of optimizing variational parameters
of quantum circuits are computationally hard in worst case
complexity [
20
]. While this is not of too much concern in
practical considerations (since it is often sufficient to find a
“good” local minimum instead of the global minimum) and
resembles an analogous situation in classical machine learning,
it does point to the fact that one should expect a rugged opti-
mization landscape, featuring different local minima as well
as saddle points. Such saddle points can indeed be a burden
to feasible and practical classical optimization of variational
quantum algorithms.
Figure 1. Stochasticity in variational quantum algorithms can help in
avoiding (strict) saddle points.
In this work, we establish the notion that in such situations,
small noise levels can actually be of substantial help. More
precisely, we show that some levels of statistical noise (specifi-
cally the kind of noise that naturally arises from a finite number
of measurements to estimate quantum expectation values) can
even be beneficial. We get inspiration from and build on a
powerful mathematical theory in classical machine learning:
there, theorems have been established that say that “noise” can
help gradient-descent optimization not to get stuck at saddle
points [
21
,
22
]. Building on such ideas we show that they can
be adapted and developed to be applicable to the variational
arXiv:2210.06723v2 [quant-ph] 8 Jun 2023