MULTI-FIDELITY MONTE CARLO: A PSEUDO-MARGINAL APPROACH 3
of low-fidelity models available. More specifically, we assume access to a sequence of low-fidelity
models that converge to a “perfect-fidelity” model in the limit. Within an MCMC algorithm, we
can approximate the perfect-fidelity target density with an unbiased estimator constructed from a
randomized truncation of the infinite telescoping series of low-fidelity target densities. This class
of multi-fidelity MCMC is an example of a pseudo-marginal MCMC (PM-MCMC) algorithm—the
unbiased estimator essentially guarantees that the algorithm is asymptotically exact in that the
limiting distribution recovers the perfect-fidelity target distribution as its marginal distribution.
Our approach introduces the fidelity of a model as an auxiliary random variable that is evolved
separately from the target variable according to its own conditional target distribution; this
technique can be used in conjunction with any suitable MCMC update that leaves the conditional
update for the target variable of interest invariant, such as M-H, slice sampling, elliptical slice
sampling, or Hamiltonian Monte Carlo. We apply the pseudo-marginal multi-fidelity MCMC
approach to several problems, including log-Gaussian Cox process modeling, Bayesian ODE
system identification, PDE-constrained optimization, and Gaussian process parameter inference.
1.1. Related work. Multi-fidelity MCMC methods are commonly applied in a two-stage proce-
dure, where the goal is to reduce the computational cost of using a single expensive high-fidelity
model by using a cheap low-fidelity model as a low-pass filter for a delayed acceptance/rejection
algorithm (Christen and Fox,2005;Efendiev et al.,2006;Cui et al.,2015); see Peherstorfer
et al. (2018) for a survey. Higdon et al. (2002) propose coupling a high-fidelity Markov chain
with a low-fidelity Markov chain via a product chain. In constrast, our approach aims to sample
from a “perfect-fidelity” target density while reducing computational cost; two-stage MCMC
algorithms result in biased estimates with respect to this target density. A related class of methods
is multilevel Monte Carlo (Giles,2008,2013;Dodwell et al.,2015;Warne et al.,2021), which uses
a hierarchy of multi-fidelity models for Monte Carlo estimation by expressing the expectation of a
high-fidelity model as a telescoping sum of low-fidelity models. Dodwell et al. (2015) use the M-H
algorithm to form the multilevel Monte Carlo estimates, simulating from a separate Markov chain
for each level of the telescoping sum. In practice multilevel Monte carlo requires choosing a finite
number of fidelities, inducing bias in the estimator with respect to the (limiting) perfect-fidelity
model. In contrast, our method uses a randomized fidelity within a single Markov chain with the
perfect-fidelity model as the target.
Our approach applies pseudo-marginal MCMC to multi-fidelity problems. There is a rich
literature developing pseudo-marginal MCMC methods (Beaumont,2003;Andrieu and Roberts,
2009) for so-called “doubly-intractable” likelihoods, which are likelihoods that are intractable to
evaluate. Several approaches in the pseudo-marginal MCMC literature are particular relevant to
our work. The first are the PM-MCMC methods introduced by Lyne et al. (2015), which describes
a class of pseudo-marginal M-H methods that use Russian roulette estimators to obtain unbiased
estimators of the likelihood. However, this method samples the variable of interest jointly with
the auxillary randomness, which often leads to sticking.
Alternatively, several methods have considered sampling the randomness separately. The idea of
clamping random numbers is explored in depth by Andrieu et al. (2010) and Murray and Graham
(2016); the latter applies to this pseudo-marginal slice sampling. In particular, our approach applies
these ideas to the specific setting of multi-fidelity models, where the random fidelity is treated as
an auxillary variable. Finally, while our approach applies to doubly-intractable problems, we are
also motivated by a larger class of multi-fidelity problems studied in the computational sciences
that may not even be inference problems, such as quantum simulations and PDE-constrained
optimization.