
Generative model for learning quantum ensemble
via optimal transport loss
Hiroyuki Tezuka∗,1,2,3, Shumpei Uno∗,2,4, and Naoki Yamamoto†,2,5
1Sony Group Corporation, 1-7-1 Konan, Minato-ku, Tokyo, 108-0075, Japan
2Quantum Computing Center, Keio University, Hiyoshi 3-14-1, Kohoku-ku, Yokohama 223-8522, Japan
3Graduate School of Science and Technology, Keio University, 3-14-1 Hiyoshi, Kohoku-ku, Yokohama, Kanagawa,
223- 8522, Japan
4Mizuho Research & Technologies, Ltd., 2-3 Kanda-Nishikicho, Chiyoda-ku, Tokyo, 101-8443, Japan
5Department of Applied Physics and Physico-Informatics, Keio University, Hiyoshi 3-14-1, Kohoku-ku, Yokohama
223-8522, Japan
Abstract
Generative modeling is an unsupervised machine learning framework, that exhibits strong performance in various
machine learning tasks. Recently we find several quantum version of generative model, some of which are even
proven to have quantum advantage. However, those methods are not directly applicable to construct a generative
model for learning a set of quantum states, i.e., ensemble. In this paper, we propose a quantum generative model
that can learn quantum ensemble, in an unsupervised machine learning framework. The key idea is to introduce
a new loss function calculated based on optimal transport loss, which have been widely used in classical machine
learning due to its several good properties; e.g., no need to ensure the common support of two ensembles. We
then give in-depth analysis on this measure, such as the scaling property of the approximation error. We also
demonstrate the generative modeling with the application to quantum anomaly detection problem, that cannot be
handled via existing methods. The proposed model paves the way for a wide application such as the health check
of quantum devices and efficient initialization of quantum computation.
1 Introduction
In the recent great progress of quantum algorithms for both noisy near-term and future fault-tolerant quantum
devices, particularly the quantum machine learning (QML) attracts huge attention. QML is largely categorised into
two regimes in view of the type of data, which can be roughly called classical data and quantum data. The former
has a conventional meaning used in the classical case; for the supervised learning scenario, e.g., a quantum system
is trained to give a prediction for a given classical data such as an image. As for the latter, on the other hand, the
task is to predict some property for a given quantum state drawn from a set of states, e.g., the phase of a many-body
quantum state, again in the supervised learning scenario. Thanks to the obvious difficulty to directly represent a huge
quantum state classically, some quantum advantage have been proven in QML for quantum data [1–3].
In the above paragraph we used the supervised learning setting to explain the difference of classical and quantum
data. But the success of unsupervised learning in classical machine learning, particularly the generative modeling,
is of course notable; actually a variety of algorithms have demonstrated strong performance in several applications,
such as image generation [4–6], molecular design [7], and anomaly detection [8]. Hence, it is quite reasonable that
several quantum unsupervised learning algorithms have been actively developed, such as quantum circuit born machine
(QCBM) [9,10], quantum generative adversarial network (QGAN) [11,12], and quantum autoencoder (QAE) [13,14].
Also, Ref. [12] studied the generative modeling problem for quantum data; the task is to construct a model quantum
system producing a set of quantum states, i.e., quantum ensemble, that approximates a given quantum ensemble. The
model quantum system contains latent variables, the change of which corresponds to the change of output quantum
state of the system. In classical case, such generative model governed by latent variables is called an implicit model. It
is known that, to efficiently train an implicit model, we are often encouraged to take the policy to minimize a distance
between the model dataset and training dataset, rather than minimizing e.g., the divergence between two probability
distributions. The optimal transport loss (OTL), which typically leads to the Wasserstein distance, is suitable for the
purpose of measuring the distance of two dataset; actually the quantum version of Wasserstein distance was proposed
in [15,16] and was applied to construct a generative model for quantum ensemble in QGAN framework [17,18].
Along this line of research, in this paper we also focus on the generative modeling problem for quantum ensemble.
We are motivated from the fact that the above-mentioned existing works employed the Wasserstein distance defined
for two mixed quantum states corresponding to the training and model quantum ensembles, where each mixed state
∗These authors equally contributed to this work.
†e-mail address: yamamoto@appi.keio.ac.jp
1
arXiv:2210.10743v1 [quant-ph] 19 Oct 2022