Generating Approximate Ground States of Molecules Using Quantum Machine Learning Jack Ceroni1 2Torin F. Stetina3 4Mária Kieferová5Carlos

2025-04-24 0 0 1.29MB 29 页 10玖币
侵权投诉
Generating Approximate Ground States of Molecules Using Quantum Machine
Learning
Jack Ceroni,
1, 2
Torin F. Stetina,
3, 4
Mária Kieferová,
5
Carlos
Ortiz Marrero,
6, 7
Juan Miguel Arrazola,
1
and Nathan Wiebe
8, 9
1
Xanadu, Toronto, ON, M5G 2C8, Canada
2
Department of Mathematics, University of Toronto, Toronto, ON, M5S 3E1, Canada
3
Simons Institute for the Theory of Computing, Berkeley, CA, 94704, USA
4
Berkeley Quantum Information and Computation Center,
University of California, Berkeley, CA 94720, USA
5
Centre for Quantum Computation and Communication Technology,
Centre for Quantum Software and Information, University of Technology Sydney, NSW 2007, Australia
6
AI & Data Analytics Division, Pacific Northwest National Laboratory, Richland, WA 99354
7
Department of Electrical & Computer Engineering,
North Carolina State University, Raleigh, NC 27607
§
8
Department of Computer Science, University of Toronto, ON M5S 1A1, Canada
9
High Performance Computing Group, Pacific Northwest National Laboratory, Richland, WA 99354
(Dated: January 3, 2023)
The potential energy surface (PES) of molecules with respect to their nuclear positions is a primary
tool in understanding chemical reactions from first principles. However, obtaining this information
is complicated by the fact that sampling a large number of ground states over a high-dimensional
PES can require a vast number of state preparations. In this work, we propose using a generative
quantum machine learning model to prepare quantum states at arbitrary points on the PES. The
model is trained using quantum data consisting of ground-state wavefunctions associated with
different classical nuclear coordinates. Our approach uses a classical neural network to convert the
nuclear coordinates of a molecule into quantum parameters of a variational quantum circuit. The
model is trained using a fidelity loss function to optimize the neural network parameters. We show
that gradient evaluation is efficient and numerically demonstrate our method’s ability to prepare
wavefunctions on the PES of hydrogen chains, water, and beryllium hydride. In all cases, we find that
a small number of training points are needed to achieve very high overlap with the groundstates in
practice. From a theoretical perspective, we further prove limitations on these protocols by showing
that if we were able to learn across an avoided crossing using a small number of samples, then
we would be able to violate Grover’s lower bound. Additionally, we prove lower bounds on the
amount of quantum data needed to learn a locally optimal neural network function using arguments
from quantum Fisher information. This work further identifies that quantum chemistry can be an
important use case for quantum machine learning.
I. INTRODUCTION
One of the most widely-studied uses of quantum com-
puters is the simulation and characterization of physical
systems. Significant effort has been dedicated to develop-
ing quantum algorithms for a variety of calculations across
many areas of physics, including condensed matter [
1
,
2
],
molecular physics [
3
,
4
], and quantum field theory [
5
,
6
].
However, in recent years, quantum chemistry has emerged
as a leading application, and considerable focus has been
placed on using quantum devices to determine the prop-
erties of molecules and materials [
7
12
]. Most proposals
for quantum computational chemistry algorithms attempt
to solve the electronic structure problem, in which the
nucleus of a molecule is assumed to be fixed, and the goal
jack.ceroni@mail.utoronto.ca
torins@berkeley.edu
maria.kieferova@uts.edu.au
§carlos.ortizmarrero@pnnl.gov
nawiebe@cs.toronto.edu
is to compute the ground-state energy of the electronic
Hamiltonian. While the general problem of computing
the ground state of an arbitrary Hamiltonian is
QMA
-
complete [
13
], under certain conditions, such as when
given access to a sufficiently high-fidelity approximation
of the true ground state [
14
16
], quantum computers can
efficiently yield accurate ground states. The most studied
proposals for solving the electronic structure problem on
circuit-based quantum computers are the variational quan-
tum eigensolver (VQE) [
17
19
] for the noisy, intermediate
scale regime, and quantum phase estimation (QPE) [
20
]
for fault-tolerant quantum computers. Both techniques
yield an approximation of the ground state energy of a
Hamiltonian, as well as the ground state itself.
While these approaches are effective in many scenarios,
their use comes with a considerable cost. For example,
for the FeMoCo molecule, a widely-used benchmark in
quantum computational chemistry, the best estimate of
QPE runtime for determining its ground state energy
is just under 4 days [
21
] on a fault-tolerant quantum
computer with millions of physical qubits. On the other
hand, VQE has the possibility of allowing for quantum
arXiv:2210.05489v3 [quant-ph] 2 Jan 2023
2
computational chemistry with fewer, noisier qubits, but
the number of measurements needed to estimate energies
is often significant, making scaling of the algorithm to
large molecules a challenge [22–25].
The challenges for both QPE and VQE worsen even
further when they are considered in the context of solving
practical problems in quantum chemistry. The electronic
structure problem assumes a fixed molecular configuration,
and therefore a fixed molecular Hamiltonian, of which
we compute the ground state. In order to determine
many dynamic or structural properties of molecules, such
as reaction barriers and optimal geometries, a molecule
must be studied in many different configurations. In gen-
eral, characterizing this behaviour requires knowledge of
a family of ground states for a set of Hamiltonians param-
eterized by classical nuclear coordinates
H
(
R
). This is an
arduous task, which requires computing many different
ground states with corresponding energies lying on a high-
dimensional potential energy surface. Running quantum
algorithms such as QPE or VQE for each configuration in-
dependently would represent a significant computational
cost even for molecules with a modest number of atoms.
We propose in this paper an alternative method for
computing ground states corresponding to a wide range
of molecular configurations, i.e., for reconstructing po-
tential energy surfaces of molecules. A key motivation
behind this work is that while the fixed nuclei electronic
structure problem is already difficult, the ultimate goal of
using quantum computers to compute accurate electronic
energies requires sampling over many different nuclear
configurations in a proposed chemical reaction coordinate.
Therefore, the generation of many ground states, and
subsequently energies and other properties, is of central
interest in taking advantage of quantum computers for
designing new materials and technologies. Instead of
computing the ground states for many discrete molecular
configurations independently, our algorithm uses a limited
collection of data and, employing techniques from ma-
chine learning, builds a model that prepares the ground
state over some region in parameter space.
Our work connects with recent progress for the task
of learning from quantum mechanical data with both
classical and quantum methods, an increasingly active
area of research. Notable results include demonstration
that classical machine learning techniques are provably
efficient for predicting and modelling certain properties
of quantum many-body systems [
26
], and that quantum
learning procedures can be more efficient than classical
learning procedures for determining specific properties of
certain unknown quantum states and processes [
26
28
].
Our proposed algorithm shares some similarities with the
above works, with the key difference being that the output
of our model is a quantum state, rather than an estimate
of some observable quantity or a classical approximation
of a quantum state [
29
,
30
]. Therefore, algorithms of
the form we proposed can be used to extract arbitrary
ground state observables, or to output states that can
be used in other quantum computational procedures re-
quiring access to ground states. From the perspective of
Ref. [
31
], this model falls into the quantum-quantum or
“QQ” category of machine learning techniques: a quantum
model trained with quantum data, and complements the
growing literature on using quantum data and quantum
machine learning to understand quantum systems (in par-
ticular, quantum chemical systems). Existing examples of
QQ machine learning include learning excited states from
ground state [
32
], compression of quantum data [
33
], and
learning of parametrized Hamiltonians [
34
], which has
been applied to spin and molecular Hamiltonians under
the name quantum meta-learning [35].
The algorithm proposed in this work is a hybrid
classical-quantum generative model, in which we train a
classical neural network to yield parameters, which when
fed into a low-depth variational quantum circuit, approxi-
mate the corresponding ground state of
H
(
R
)for a range
of values of
R
. To train our model, we assume access to
quantum data: ground states of
H
(
R
)for a collection of
coordinates
{Ri}N
i=1
, which can be loaded into a quantum
computer. Since ground state preparation is a resource
intensive task, the amount of quantum data needed to
learn a model is a key metric for quantifying the efficiency
and the feasibility of the algorithm. Ideally, a model of
this form should generalize to new values of
R
not con-
tained in the training data. This allows us to generate
approximations of previously unseen molecular ground
states, at the more modest price of executing a shallow,
variational quantum circuit for some set of parameters
determined by a classical neural network.
To test these capabilities, we perform extensive numeri-
cal experiments for a collection of different molecules, and
find that even with few data points, there is good gen-
eralization to unseen geometries in the potential energy
surface. Ultimately, the aim of our proposal is to provide
a concrete first step towards the development of practi-
cal techniques based on quantum machine learning for
alleviating the cost of computing ground states of a param-
eterized molecular Hamiltonian. Under this framework,
hard-to-obtain quantum data, originating from quantum
algorithms or physical experiments [
36
], is the resource
that we attempt to leverage.
It is important to note that the particular generative
model proposed is only one member of a large family
of quantum-classical machine learning architecture for
preparing ground states. Advanced models may utilize
more sophisticated choices of cost function, classical neu-
ral network architecture, initialization, and circuit con-
struction than those considered in this paper. The goal
of this work is to both discuss the general concept of
generative quantum machine learning applied to quantum
chemistry, as well as provide a concrete example of what
such a model would look like. As a result, we explore
both practicalities associated with the particular model,
including gradient sample complexity and numerical ex-
periments, as well as more general considerations about
abstract quantum state-learning procedures.
We begin in Sec. II by outlining a general architecture
3
and training strategy for a generative model which pre-
pares ground states of a parameterized Hamiltonian. In
Sec. III, we discuss the details of training, and provide
estimates on the sample complexity required for com-
puting gradients of the model, which is necessary for
optimization. In Sec. IV, to better understand the limits
of quantum generative models, we introduce theoretical
bounds related to data complexity of general quantum
state-learning algorithms, and interpret these results in
the context of our generative model. We conclude in
Sec. V by providing numerics to support the quality of
our model, demonstrating that one can effectively learn
out-of-distribution ground states, and thus the resulting
potential energy surfaces, of the H
2
, H
+
3
, H
4
, BeH
2
, and
H2O molecules, to a high degree of accuracy.
II. A GENERATIVE MODEL FOR PREPARING
ELECTRONIC GROUND STATES
Quantum chemistry has, in recent years, become
a major focus for the development of quantum algo-
rithms [
10
,
37
,
38
]. The specific problem that has domi-
nated most discussion surrounding quantum computing
in this space is the electronic structure problem, which
involves finding the minimum energy configuration for a
molecule. This problem reduces to the problem of com-
puting the groundstate energy of a quantum system. The
electronic structure problem in a fixed basis is, in general,
QMA
-Hard [
39
] Regardless, in many cases though a state
with sufficient overlap can be identified and the overlap
with the state can be computed within constant error
using a polynomial number of quantum gates [
23
,
40
,
41
].
The Hamiltonian for the electrons within the Born-
Oppenheimer approximation (which states that the nu-
clear motion is uncoupled with the electronic motion) can
be written as
H(R) = X
pq
hpq(R)a
paq+1
2X
pqrs
hpqrs(R)a
pa
qaras,(1)
where
a
p
and
ap
are the fermionic creation and annihi-
lation operators acting on the p-th orbital, and
hpq(R) = Zdr φ
p(r) 2
2X
I
ZI
|rRI|!φq(r),
(2)
hpqrs(R) = Zdr1dr2
φ
p(r1)φ
q(r2)φr(r2)φs(r1)
|r1r2|,(3)
are the one and two-electron integrals in the molecular
orbital basis,
φp
(
r
), yielded from the Hartree-Fock opti-
mization procedure. The one and two-electron integrals,
and subsequently the molecular orbitals depend implicitly
on
R
, which is the general nuclear coordinate associated
with the fixed nuclear configuration in 3-dimensional space.
We then utilize the Jordan-Wigner transform in order to
map the fermionic creation and annihilation operators to
qubit operators, which can be implemented in a quantum
circuit [42, 43].
The best known costs for estimating the groundstate
energy at a specific nucelar configuration scales as [
10
]
˜
O
(
Nλ/
)where
N
is the number of spin-orbitals used
in the model and
λ
is a quantity that depends on the
details of the Hamiltonian and the basis set chosen. For
the case where a planewave basis is chosen an explicit
scaling can be achieved of
˜
O
(
N3/
)[
38
]; however, such
an explicit scaling is difficult to derive in a Gaussian
basis (like those commonly used in chemistry) but can be
numerically verified to scale polynomially in
N
for most
examples considered.
The challenge behind all of this is that the cost of
preparing these groundstates is difficult. It requires per-
forming phase estimation to learn the groundstate en-
ergy within the eigenvalue gap, which leads to a cost of
O
(
N3/|E1E0|
)for the case of planewaves. Such costs
for generating a quantum state at any point on the poten-
tial energy surface can be challenging for problems with
small gaps and so a question that naturally emerges is
whether there exist faster methods for preparing approx-
imate groundstates on quantum computers. In general,
we do not expect this to be possible without prior infor-
mation; however, when approaching chemistry we almost
never have a uniform prior. We are frequently met with
a host of prior experiments and often have systems that
are smooth enough so nearby queries yield information
about the quantum state in question. Our aim is to use
quantum machine learning as a paradigm to estimate
new states on the potential energy surface of a molecule.
We discuss how generative quantum machine learning to
approach this challenge below.
A. Generative QML Models for Chemistry
The generative model that we propose contains both a
classical and quantum component. We begin by fixing a
parameterized quantum circuit
U
(
θ
). Although the struc-
ture of
U
can vary, we assume that it can be expressed
as a series of parameterized rotations [17],
U(θ) =
NP
Y
j=1
ejHj,(4)
where each
Hj
is Hermitian and chosen prior to learning.
Given a parameterized Hamiltonian
H
(
R
), the goal is
to learn a map
R7→ θ
(
R
), such that
U
(
θ
(
R
))
|ψinit
(
R
)
i
approximates the ground state
|ψ0
(
R
)
i
of
H
(
R
)over some
range of
R
, where
|ψinit
(
R
)
i
is some pre-chosen initial
state. For the sake of simplicity,
H
(
R
)is taken to be
non-degenerate over
R
in the region considered. In order
4
FIG. 1: A diagram showing how a neural network is used to parameterize a map from coordinates Rto a state
|ψ(ν(R;γ))i. The goal is to find the optimal parameters γ=γsuch that when they are used in the generative
procedure, the resulting map from
R
to
|ψ
(
ν
(
R
;
γ
))
i
approximates the map from
R
to the ground state
|ψ0
(
R
)
i
. This
allows for accurate extraction of ground state observables, such as ground state energies, from our model, over R.
to find the mapping
R7→ θ
(
R
), we are given access to a
collection of quantum data, D, of the form
D=
Ri,
mi
z }| {
|ψ0(Ri)i, . . . , |ψ0(Ri)i
N
i=1
,(5)
where each
Ri
is unique and we have access to
mi
copies of
a state
|ψ0
(
Ri
)
i
. We assume a polynomially large number
of data, which should be contrasted with the exponentially
large number of possible grid points
Ri
that define a full
potential energy surface in high dimensions. We do not
explicitly consider the origin of the data set, but it can
be created with several different methods, possibly by re-
peated application of QPE or VQE for the parameters
Ri
,
or even by transduction of quantum states yielded from
an experiment into a quantum device, as is envisioned
in Ref. [
36
]. We note that the model is not restricted
to returning parameterized ground states: given training
data, it can be used to construct an approximation of
any parameterized state. For example, one could imag-
ine using a similar procedure to produce a model of a
parameterized excited state
|ψj
(
R
)
i
. We focus on ground
state generation for the sake of concreteness, and due to
the fact that computing ground states is a problem of
practical importance.
Let
ν
(
R
;
γ
)denote a classical neural network, where
γ
are the trainable parameters, which induces a func-
tion
R7→ ν
(
R
;
γ
). Using the data
D
along with
ν
,
we perform an optimization procedure which yields
γ
such that the output state
U
(
ν
(
Ri
;
γ
))
|ψinit
(
Ri
)
i
approx-
imates
|ψ0
(
Ri
)
i
, for each
Ri
in the training data. Define
θ
(
R
) :=
ν
(
R
;
γ
)to be the output of the model. The goal
of such a training procedure is good generalization for all
values of R. In other words,
|ψ(ν(R;γ))i:= U(ν(R;γ))|ψinit(R)i
=U(θ(R))|ψinit(R)i,(6)
should approximate
|ψ0
(
R
)
i
for a wider range of
R
outside
of the training set. To find the optimal parameters
γ
, we
minimize a cost function
C
(
γ
)over the training data. In
this work, we choose
C
to be the average infidelity of the
state produced by the model with each unique training
example (each state in
D
corresponds to a unique
Ri
).
More specifically, the cost function is
C(γ) = 1 1
N
N
X
i=1 |hψ0(Ri)|ψ(ν(Ri;γ))i|2,(7)
where for sufficiently expressive
U
, minimization of
C
will
result in
γ
such that
|ψ
(
ν
(
Ri
;
γ
))
i≈|ψ0
(
Ri
)
i
for each
Ri
. One immediate benefit of using the infidelity cost
function is that it admits relatively simple gradients of
the form
C(γ)
γk
=1
N
N
X
i=1 2Re [hψ0(Ri)|ψ(ν(Ri;γ))i]
NP
X
a=1
ν(Ri;γ)a
γkDψ0(Ri)ψ(θ)
θaEθ=ν(Ri;γ)!,(8)
where
|ψ
(
θ
)
i
:=
U
(
θ
)
|ψinit
(
Ra
)
i
for each term in the sum
and
NP
is the output dimension of the classical neural
network, i.e, the number of parameters of the quantum
circuit (for a derivation, see Sec. III). Overlaps between
the training states, and the output/output derivatives of
the parameterized circuit
U
can be computed via sam-
pling, or amplitude estimation and the linear combination
of unitaries (LCU) method [
37
,
44
], while the derivatives
5
of
ν
can be computed efficiently in the case of deep neural
networks via backpropagation [
45
]. With a procedure for
computing derivatives, it is then possible to minimize
C
via gradient-based optimization methods, such as gradient
descent or Adam [
46
]. The computational cost of comput-
ing gradients of the cost function in Eq.
(7)
is discussed
in detail in Section III.
We conclude this section by discussing one of the central
challenges in designing quantum machine learning models:
barren plateaus, and how the proposed model may be
robust to this issue. The barren plateaus problem is the
observation that many classes of parameterized quantum
circuits suffer from exponentially vanishing gradients over
large parts of the parameter space [
47
]. Barren plateaus
highlight the necessity to incorporate inductive biases into
parameterized quantum circuits based on the structure
of a particular problem, by choosing appropriate circuit
ansatzes, cost function, and initialization, among other
hyperparameters. As is discussed in Refs. [
47
,
48
],
C
(
γ
)of
the general form in Eq.
(8)
is conducive to barren plateaus.
This issue can, in some cases, be resolved by choosing a
more sophisticated cost function [
48
,
49
]. However, before
swapping
C
for one of these functions, it is important to
note that any generative model tasked with returning a
molecular ground state can have as its initial guess an
approximation of the true ground state. In our work we
set
|ψinit
(
R
)
i
:=
|ψHF
(
R
)
i
, where
|ψHF
(
R
)
i
is the Hartree-
Fock state at geometry
R
; an approximate solution to the
electronic Schrodinger equation which can be computed
efficiently on a classical device.
The parameterized quantum circuit
U
can then be
thought of as applying corrections to the Hartree-Fock
state. This choice makes the initialization of our model
far from random. In addition, since we have access to
the Hamiltonian
H
(
R
)of which we are attempting to
prepare the ground state, we can use this knowledge
to tailor our circuit ansatz to each particular molecule
considered, using the ADAPT-VQE algorithm [
50
] (see
Sec. V for details). Ref. [
51
] provides empirical evidence
that in many cases, adaptive circuits do not suffer from
the issue of barren plateaus. It is therefore reasonable
to hypothesize that initialization in the Hartree-Fock
state and an adaptively-prepared circuit will, in many
cases, constrain the model’s optimization to a region that
does not suffer from the barren plateau problem. This
conclusion is supported in the numerics (Sec. V), where
we observe good performance when using the infidelity
cost function of Eq. (7) for training.
III. GRADIENT SAMPLING COMPLEXITY
While the model described in Section II is a hybrid
quantum-classical procedure, the part of the algorithm
that is executed on a quantum device can be reduced to
the calculation of gradients of the cost function
C
(
γ
). As
a result, understanding the quantum sample complexity
of gradient calculations allows us to better understand the
cost of running the entire algorithm. The other important
aspect is the number of training steps required to find
the optimal
γ
, but this task is much more challenging,
and likely varies considerably on a case-by-case basis.
In this section, we discuss the sample complexity re-
quired to compute gradients of the average infidelity cost
function introduced in Section II. Recall that
|ψ
(
θ
)
i
:=
U
(
θ
)
|ψinit
(
Ri
)
i
for some
Ri
. We can express the gradient
of the i-th overlap terms in Eq. (7) as
γk|hψ0(Ri)|ψ(ν(Ri;γ))i|2
=
γkhψ0(Ri)|ψ(ν(Ri;γ))ihψ(ν(Ri;γ))|ψ0(Ri)i
=hψ(ν(Ri;γ))|ψ0(Ri)iDψ0(Ri)ψ(ν(Ri;γ))
γkE+h.c.
= 2Re hψ(ν(Ri;γ))|ψ0(Ri)iDψ0(Ri)ψ(ν(Ri;γ))
γkE
2Re"hψ(ν(Ri;γ))|ψ0(Ri)i
NP
X
a=1
ν(Ri;γ)a
γk
×Dψ0(Ri)ψ(θ)
θaEθ=ν(Ri;γ)#.(9)
Now, note that
ψ(θ)
θa=
θa
U(θ)|ψinit(Ri)i
=
θa"NP
Y
j=1
ejHj#|ψinit(Ri)i
=i"Y
j<a
ejHjHaY
ja
ejHj#|ψinit(Ri)i
=i"Y
j<a
ejHjHa Y
j<a
ejHj!
×Y
j<a
ejHjY
ja
ejHj#|ψinit(Ri)i
=i"Y
j<a
ejHjHa Y
j<a
ejHj!#|ψ(θ)i
=iˆ
Ha(θ)|ψ(θ)i,(10)
where
ˆ
Ha
(
θ
) :=
Qj<a ejHjHaQj<a ejHj
. Let
ˆ
Ha
:=
ˆ
Ha
(
ν
(
Ri
;
γ
)). For brevity, we will henceforth refer
to
|ψ
(
ν
(
Ri
;
γ
))
i
as
|ψi
, and
|ψ0
(
Ri
)
i
as
|ψii
. Using Eq.
(9)
and Eq. (10), we get
摘要:

GeneratingApproximateGroundStatesofMoleculesUsingQuantumMachineLearningJackCeroni,1,2TorinF.Stetina,3,4MáriaKieferová,5CarlosOrtizMarrero,6,7JuanMiguelArrazola,1andNathanWiebe8,91Xanadu,Toronto,ON,M5G2C8,Canada2DepartmentofMathematics,UniversityofToronto,Toronto,ON,M5S3E1,Canada3SimonsInstitutefort...

展开>> 收起<<
Generating Approximate Ground States of Molecules Using Quantum Machine Learning Jack Ceroni1 2Torin F. Stetina3 4Mária Kieferová5Carlos.pdf

共29页,预览5页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:29 页 大小:1.29MB 格式:PDF 时间:2025-04-24

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 29
客服
关注