
2 MAX NENDEL AND ALESSANDRO SGARABOTTOLO
reference distribution µmay be available. This is a special instance of model uncertainty
appearing, for example, in the context of catastrophic risk in reinsurance or default risk
within large credit portfolios in banking.
In the economic literature, model uncertainty is also referred to as Knightian uncertainty,
and a standard way to deal with it is to look at worst case losses among a set of plausible
probability distributions. In our study, we follow this approach, and estimate worst case
losses over the set Ppof all Borel probability measures on Hwith finite moment of order
p∈(1,∞), weighting the different measures via a penalization term depending on the
p-Wasserstein distance from a reference model µ(which is assumed to have finite moment
of order pas well). This leads to an expression of the form
If:= sup
ν∈PpZH
f(z)ν(dz)−φWp(µ, ν).(1.1)
Here, the penalty function φ: [0,∞)→[0,∞], which is assumed to be nondecreasing with
φ(0) = 0, reflects a degree of confidence in the reference measure, which could, for example,
be related to the availability of data for the estimation of µ. The value φ(a) = ∞for some
a > 0corresponds to a rejection of every model νwith Wasserstein distance Wp(µ, ν)≥a,
and the limit case φ=∞ · (0,∞)resembles perfect confidence in the measure µ.
Functionals of the form (1.1) belong to the class of convex risk measures and, under
suitable conditions on the penalty function φ, to the class of coherent risk measures, cf.
[2] and [18]. Moreover, they are widely studied in the context of distributionally robust
optimization problems, see, for example, [5,20,26,31,33,34], where the authors usually
consider an additional optimization procedure, leading to an inf-sup-formulation.
A standard approach to tackle the infinite-dimensional optimization related to (1.1) is to
look for a suitable dual formulation, for example, by transforming the primal problem into
a superhedging problem. For example, in [5], the authors transform a class of robust opti-
mized certainty equivalents (OCEs) into a one-dimensional optimization that leads to an
explicit correction term. In general, however, this approach leads to a nested optimization
problem, which can be numerically challenging.
We therefore look at this problem from a similar yet different angle, and aim to identify
a parametric version of the functional (1.1) together with suitable optimizing directions.
This idea is merely related to the paradigm of looking for extreme points in Wasserstein
balls; a topic that has been explored in detail in the case where the reference measure is an
empirical distribution (uniform over the samples) or, more generally, a convex combination
of Dirac measures. In [33], it is shown that extreme points of Wasserstein balls centered
in a measure supported on at most npoints are supported on at most n+ 3 points. The
paper [29] refines this result, showing that these extreme distributions are in fact supported
on n+ 2 points. Finally, in [26], the authors show that, under stronger assumptions on
the loss function, the infinite-dimensional optimization problem can be solved via a convex
shifting of the support points and that the optimizing distribution is supported on the same
number of points as the reference measure. In the case, where µis a convex combination
of Dirac measures, we get a similar result, but in a different fashion, cf. Section 3.1.
The key idea of our approach is to look for a parametric version of the functional (1.1)
in terms of a first order approximation as the level of uncertainty related to the penalty
function φtends to zero. More precisely, we introduce a scaling parameter h > 0, and
substitute φwith the rescaled version φh:= hφ(·/h), which allows to control the level of
uncertainty in terms of h. We then consider the operator I(h), given by
I(h)f:= sup
ν∈PpZH
f(z)ν(dz)−φhWp(µ, ν)(1.2)