
Petrus Mikkola, Julien Martinelli, Louis Filstroff, Samuel Kaski
Definition 1
(BO regret)
.
The regret of the BO algorithm
that spends
Λcost
and returns the final choice
xchoice
, is de-
fined by
R(Λcost,xchoice) := (f∗−f(m)(xchoice)if Λcost ≤Λ,
∞otherwise
where
f∗= maxx∈X f(m)
is the global maximum of the
primary IS.
Definition 2
(Number of queries)
.
Let
T:= bΛ/λmc
be
the available number of primary IS queries. Let
T(`)
be the
random variable describing the number of
`th
-IS queries
spent by the MFBO algorithm.
There are two popular choices for
xchoice
. First, the Bayes-
optimal choice
xchoice = arg max
x∈X
µ(x, m),
where
µ(x, m)
is the posterior mean of the GP model given
all the data up to the final query
Tlast =Pm
`=1 T(`)
. The
regret in this case is called the inference regret. Second, the
simple choice
xchoice = arg max
t∈JT(m)K
f(m)(xt),
where
(x1, ..., xT(m))
is the primary IS acquisition trajec-
tory returned by the MFBO algorithm. The regret in that
case is called the simple regret.
Lastly, we provide an informal definition for an unreliable
IS. To the best of our knowledge, no formal definition has
been proposed in the literature so far, in view of the non-
trivial character of the task. Broadly speaking, an auxiliary
IS
`
is said to be unreliable, if querying
f(`)(x)
does not
lead to a decreasing inference regret when averaged over all
sequences of queries
x∈ X
and all datasets
D
of any size.
As such, the relevance on an IS then also depends on the
MOGP model chosen.
3 RELATED WORK
As discussed in the introduction, many different multi-
fidelity extensions of Bayesian optimization have been pro-
posed in the literature; we refer the interested reader to
Takeno et al. (2020, Section 5) for a review. The closest to
our work are methods that do not assume a hierarchy be-
tween the sources (e.g., when the degree of fidelity cannot
be assessed in advance), as by Lam et al. (2015), where the
focus lies in designing a GP model that takes into account
the non-hierarchical nature of the sources. The multi-fidelity
kernel introduced by Poloczek et al. (2017) (see Supplemen-
tary F) is one example of such a design.
Surprisingly, the problem of the potential performance
degradation of MFBO algorithms has been largely ignored
in the literature, with the exception of Kandasamy et al.
(2016), who noticed that their multi-fidelity method per-
formed poorly compared to all single-fidelity variants in one
experiment (Kandasamy et al., 2016, Supplementary D.3).
Lastly, we mention that robustness has been studied for
vanilla BO in the context of (sometimes adversarially) noisy
inputs or outputs (Martinez-Cantin et al., 2018; Bogunovic
et al., 2018; Fr
¨
ohlich et al., 2020; Kirschner and Krause,
2021). This notion of robustness is fundamentally different
from ours, indeed, we wish to provide guarantees that the
addition of an auxiliary IS (or several) will not lead to worse
performance w.r.t. vanilla BO.
4 PITFALLS OF MFBO METHODS
We now demonstrate, on a simple example, the influence
of the auxiliary IS quality on the performance of MFBO
algorithms. Let us consider the Hartmann6D function as the
objective (i.e., the primary IS). We examine two scenarios:
in the first one, the auxiliary IS is informative, consisting of
a biased version of the primary IS, with a degree of fidelity
l= 0.2
. In the second scenario, the auxiliary IS is taken
to be the 6-dimensional Rosenbrock function, an irrelevant
source for this problem. Analytical forms for these examples
can be found in Supplementary G. We evaluate the multi-
fidelity maximum-entropy search (MF-MES) method from
Takeno et al. (2020) on these two scenarios, as well as
its single-fidelity counterpart (SF-MES), and our proposed
algorithm, rMFBO, built on top of these methods (rMF-
MES). In both cases, the cost of the primary IS is set to 1,
and cost of the auxiliary IS to 0.2. The simple regret of the
three algorithms is displayed in Figure 1.
When the auxiliary IS is informative (left panel), MF-MES
converges faster than SF-MES. This is the expected behavior
from MFBO algorithms: they use cheap IS queries in the
beginning to clear out unpromising regions of the space at a
low cost, which eventually speeds up convergence. However,
when the auxiliary IS is irrelevant (right panel), there is a
clear gap between MF-MES and SF-MES, even in the long
run. This demonstrates the inability of MF-MES to deal
with an irrelevant IS. In that scenario, we hypothesize that
the budget is wasted on uninformative queries, and thus too
many rounds are spent on learning that the sources are not
correlated (Figure 1, right bar plot), leading to a sub-optimal
data acquisition trajectory.
There is therefore a need for a robust method for such a
scenario. This is what the proposed rMF-MES, formally
introduced in the next section, achieves, by taking the best of
both worlds: sticking close to the single fidelity track in case
of an irrelevant IS, while using informative lower-fidelity
queries to accelerate convergence.