
In Section 4we develop a calculus for anisotropic smoothness and provide examples.
In Section 5we analyse the algorithm.
In Section 6we provide an equivalent interpretation in terms of a difference of Φ-convex
approach.
In Section 7we consider applications based on the examples developed in Section 4.
2 Notation and preliminaries
We denote by ⟨·,·⟩ the standard Euclidean inner product on Rnand by ∥x∥:= p⟨x, x⟩for
any x∈Rnthe standard Euclidean norm on Rn. In accordance with [36,42] we extend the
classical arithmetic on Rto the extended real line R:= R∪ {−∞,+∞}. We define upper
addition −∞ ˙
+∞=∞and lower addition −∞+
.∞=−∞, and accordingly upper subtraction
∞˙
− ∞ =∞and lower subtraction ∞ −
.∞=−∞. The effective domain of an extended
real-valued function f:Rn→Ris denoted by dom f:= {x∈Rn:f(x)<∞}, and we say
that fis proper if dom f̸=∅and f(x)>−∞ for all x∈Rn; lower semicontinuous (lsc) if
f(¯x)≤lim infx→¯xf(x) for all ¯x∈Rn; super-coercive if f(x)/∥x∥→∞as ∥x∥→∞. We
define by Γ0(Rn) the class of all proper, lsc convex functions f:Rn→R. For any functions
f:Rn→Rand g:Rn→Rwe define the infimal convolution or epi-addition of fand gas
(f□g)(x) = infy∈Rng(x−y) + f(y). For a proper function f:Rn→Rand λ≥0 we define
the epi-scaling (λ ⋆ f)(x) = λf(λ−1x) for λ > 0 and (λ ⋆ f)(x) = δ{0}(x) otherwise. We adopt
the operator precedence for epi-multiplication and epi-addition from the pointwise case. We
denote by Rn
+= [0,+∞)nthe nonnegative orthant and by Rn
++ = (0,+∞)nthe positive
orthant. Let g−:= g(−(·)) denote the reflection of g. Since convex conjugation and reflection
commute we adopt the notation: (g−)∗= (g∗)−=: g∗
−. We adopt the notions of essential
smoothness, essential strict convexity and Legendre functions from [40, Section 26]: We say
that a function f∈Γ0(Rn) is essentially smooth, if int(dom f)̸=∅and fis differentiable
on int(dom f) such that ∥∇f(xν)∥ → ∞, whenever int(dom f)∋xν→x∈bdry dom f,
and essentially strictly convex, if fis strictly convex on every convex subset of dom ∂f,
and Legendre, if fis both essentially smooth and essentially strictly convex. We denote by
N0:= N∪ {0}. Otherwise we adopt the notation from [42].
Our concepts and algorithm involve a reference function ϕwhich complies with the fol-
lowing standing requirement which is assumed valid across the entire paper unless stated
otherwise:
(A1) ϕ∈Γ0(Rn) is Legendre with dom ϕ=Rn.
The dual reference function ϕ∗is Legendre as well [40, Theorem 26.3] but does not necessarily
have full domain. Instead, thanks to a conjugate duality between super-coercivity and full do-
main [6, Proposition 2.16], ϕ∗is super-coercive. The gradient mapping ∇ϕ:Rn→int dom ϕ∗
is a diffeomorphism between Rnand int dom ϕ∗with inverse (∇ϕ)−1=∇ϕ∗[40, Theorem
26.5]. In fact, super-coercivity is a standard assumption for the mirror-descent and Bregman
(proximal) gradient algorithm as it ensures well-definedness of the mirror update [5, Lemma
2(ii)]. Since the framework that will be developed in this paper corresponds to a certain dual
Bregman approach the full-domain restriction is somewhat natural.
The running example considered in this paper is the exponential reference function:
Example 2.1.Let ϕ:Rn→Rdefined by ϕ(x) = SumExp(x) := Pn
j=1 exp(xj). Then
ϕ∈Γ0(Rn) is Legendre with full domain and thus complies with our standing assumption.
4