LINEAR INVERSE PROBLEMS WITH HESSIAN-SCHATTEN TOTAL VARIATION 2
1. Introduction
Broadly speaking, the goal of an inverse problem is to reconstruct an unknown signal
of interest from a collection of (possibly noisy) observations. Linear inverse problems, in
particular, are prevalent in various areas of signal processing, such as denoising, impaint-
ing, and image reconstruction. They are defined via the specification of three principal
components: (i) a hypothesis space Ffrom which we aim to reconstruct the unknown
signal f∗∈ F; (ii) a linear forward operator ν:F → RMthat models the data acquisition
process; and, (iii) the observed data that is stored in an array y∈RMwith the implicit
assumption that y≈ν(f∗). The task is then to (approximately) reconstruct the unknown
signal f∗from the observed data y. From a variational perspective, the problem can be
formulated as a minimization of the form
f∗∈arg min
f∈F
(E(ν(f),y) + λR(f)) ,(1)
where E:RM×RM→Ris a convex loss function that measures the data discrepancy, R:
F → Ris the regularization functional that enforces prior knowledge on the reconstructed
signal, and λ > 0 is a tunable parameter that adjusts the two terms.
The use of regularization for solving inverse problems dates back to the 1960s, when
Tikhonov proposed a quadratic (`2-type) functional for solving finite-dimensional prob-
lems [Tik63]. More recently, Tikhonov regularization has been outperformed by `1-type
functionals in various settings [Tib96,DE03]. This is largely due to the sparsity-promoting
effect of the latter, in the sense that the solution of an `1-regularized inverse problem can be
typically written as the linear combination of a few predefined elements, known as atoms
[Don06b,BDE09]. Sparsity is a pivotal concept in modern signal processing and consti-
tutes the core of many celebrated methods. The most notable example is the framework
of compressed sensing [CRT06,Don06a,EK12], which has brought lots of attention in the
past decades.
In general, regularization enhances the stability of the problem and alleviates its inherent
ill-posedness, especially when the hypothesis space is much larger than M. While this can
happen in the discrete setting (e.g. when F=Rdwith dM), it is inevitable in the con-
tinuum where Fis an infinite-dimensional space of functions. Since naturally occurring
signals and images are usually indexed over the whole continuum, studying continuous-
domain problems is, therefore, undeniably important. It thus comes with no surprise to
see the rich literature on this class of optimization problems. Among the classical examples
are the smoothing splines for interpolation [Sch88,Rei67] and the celebrated framework of
learning over reproducing kernel Hilbert spaces [Wah90,SHS01]. Remarkably, the latter
laid the foundation of numerous kernel-based machine learning schemes such as support-
vector machines [EPP00]. The key theoretical result of these frameworks is a “representer
theorem” that provides a parametric form for their optimal solutions. While these ex-
amples formulate optimization problems over Hilbert spaces, the representer theorem has
been recently extended to cover generic convex optimization problems over Banach spaces
[BCDC+19,BC20,Uns21,UA22]. In simple terms, these abstract results characterize the
solution set of (1) in terms of the extreme points of the unit ball of the regularization