
𝑓(𝒙1), 𝑓 (𝒙2), ..., 𝑓 (𝒙𝑁), from which the sample mean and sample variance are calculated. Monte Carlo methods are
known to converge as (𝑁−1
2)which, if 𝑓is expensive to evaluate, can make such methods computationally expensive or
even prohibitive.
The slow convergence of the errors with simple MC can in some cases be bypassed by using surrogate models [11].
With this approach, the 𝑁input-output samples of 𝑓are used to train a surrogate model 𝑓(𝑁). Then 𝑓(𝑁)is evaluated
𝑀times, and the outputs are used to estimate the mean and variance. The advantage of this method is that 𝑓(𝑁)is
computationally cheaper than 𝑓, and evaluating it 𝑀times with 𝑀 ≫ 𝑁 has negligible runtime. The bottleneck of this
method is in obtaining the 𝑁samples for the training set. If 𝑓is low-dimensional such that 𝑁 ≫ 𝑑, the surrogate model
is likely to have a small bias, however in high-dimensional cases with 𝑁 < 𝑑 the surrogate will be biased (see the curse
of dimensionality [13,14]), making also the mean and variance estimations biased. The specific case where 𝑁 < 𝑑, and
where increasing 𝑁is not possible or computationally expensive, is the main focus of this paper. The aim is to find a
method more accurate than simple MC for a fixed computational budget 𝑁.
In this regard, multifidelity Monte Carlo (MFMC) [15,16,17,18] offers a promising approach. Multifidelity methods
combine low and high fidelity models to accelerate the solution of an outer-loop application, like uncertainty quantific-
ation, sensitivity analysis or optimisation. The low fidelity models, that are either simplified models, projection-based
models or data-fit surrogates, are used for speed up, while the high fidelity models are kept in the loop for accuracy and/or
convergence. Thus MFMC offers a framework to combine a biased surrogate model 𝑓(𝑛)with the high fidelity model 𝑓,
in such a manner that unbiased estimates of the mean and variance are computed. The crucial point with MFMC is then,
for a given number of high fidelity evaluations 𝑁, optimising the trade-off between how many of them are used in the
multifidelity estimators and how many for training the surrogate. This optimisation problem has been recently addressed
in [19]. Nevertheless, there is no guarantee that, for a given 𝑁, the MFMC estimates will be more accurate than simple
MC, especially in high-dimensional cases with 𝑁 < 𝑑, where 𝑓(𝑛)is likely to have a large bias.
To address the challenges presented by high-dimensional UQ in existing approaches, we introduce the Lasso Monte
Carlo (LMC) method, a variation on MFMC. With LMC we propose a new data management strategy in which the high
fidelity samples are reused several times both to train multiple surrogates and in the multifidelity estimators. The resulting
algorithm is such that the estimations are guaranteed to be equally or more accurate than simple MC and MFMC, under
certain assumptions. This new approach can be viewed as a variance reduction technique, based on a two-level version
of MFMC, and Lasso regression (least absolute shrinkage and selection operator), simultaneously performing regression
analysis, variable selection and regularization [20].
It is worth mentioning the relation between multifidelity, multilevel, and multi-indexing methods: Multilevel Monte
Carlo (MLMC) [21] is a special case within the broader framework of multifidelity methods, in which typically a hierarchy
of low fidelity models is derived by varying a parameter (e.g. mesh width). A generalization of MLMC is introduced in
[22], so called multi-index Monte Carlo methods, which use multidimensional levels and high-order mixed differences to
reduce the computational cost and the variance of the resulting estimator.
The remainder of the paper is organised as follows: we start with a review of MFMC for the estimation of central
moments, and in particular we derive the expressions for the two-level estimators. This is followed by a discussion on the
trade-off between accuracy and computational costs of MFMC. The new method LMC is then introduced, and we prove
that it is equally or more accurate than simple MC. We then review the theory behind the Lasso regression method, and
show how it can be used in the LMC algorithm. Finally, in section 3LMC is benchmarked on a variety of examples. Proofs
for all the theorems and lemmas are provided in appendix A.
1.1 Notation & Assumptions
Throughout the paper, bold letters represent vectors 𝒙= (𝑥1, 𝑥2, ..., 𝑥𝑑), with 𝑑the dimension. A lower case letter 𝑥
is a realisation of a random variable 𝑋, and in the case where 𝑋is a multivariate random variable of dimension 𝑑, a
2