Generalized Bayes Approach to Inverse Problems with Model Misspecification 2
1.1. Contributions
The main contributions of this paper are the following:
1 A theory of model comparison for Gibbs posteriors that enables model comparison for
loss functions. We define a notion of “predictive performance” for Gibbs posterior and
study its theoretical properties.
2 We develop a particle filter and importance sampling method to simultaneously sample
from the underlying Gibbs posterior and calibrate the regularization parameter that
balances the loss function and regularization with respect to the prior. Our calibration
procedure minimizes a novel leave-one-out cross-validation (LOOCV) objective. Due
to the distributional nature of the solution, existing cross-validation algorithms are not
immediately applicable.
3 We prove the stability and consistency of Gibbs posteriors. We show the continuity
of Gibbs posterior as a mapping of the data in various distances for probability
distributions. Our proposed upper bound improves on existing upper bounds that are
vacuous when the perturbation to the data is large. We also study the asymptotic
behavior of Gibbs posteriors in the large sample limit. The technical aspects of a
consistency proof rely on tools in the robust Bayes estimation literature. We also study
the asymptotics of a predictive distribution used for model selection associated with the
Gibbs posterior.
1.2. Prior work
The relation between the regularized least-squares problem proposed by Tikhonov and
Arsenin [1977] and the maximum a posteriori (MAP) estimation problem in Bayesian
statistics has been known for some time. Bayesian methods for inverse problems have been
successfully adopted in diverse domains, nicely summarized by Kaipio and Somersalo [2005].
Recent literature [Cotter et al.,2009,Stuart,2010,Cotter et al.,2013] has extended the
Bayesian framework with Gaussian likelihood to infinite-dimensional settings. The Gibbs
posterior framework [Bissiri et al.,2016,Jiang and Tanner,2008,Martin et al.,2017] is not
new, and its application in inverse problems was studied by Zou et al. [2019], Dunlop and
Yang [2021]. Similar concepts have been studied by Gr¨unwald and Langford [2007], Gr¨unwald
and van Ommen [2017], Miller and Dunson [2019], Bhattacharya et al. [2019], among others,
for improving the robustness of Bayesian inference under model misspecification. The novel
model selection theory we develop in this paper can be viewed as an analog of the theory
of Bayesian model selection and Bayesian cross-validation under model misspecification
[Bernardo and Smith,2009]. Computationally, we rely on sequential Monte Carlo and
particle filters algorithms. These algorithms have gained recent attention for potential use in
Bayesian inverse problems. [Kantas et al.,2014,Beskos et al.,2015] have used particle filters
to solve parabolic and elliptic inverse problems. Zou et al. [2019] have proposed a combination
of particle filter and reduced order models for improved computational efficiency.
A vast amount of literature exists on quantifying uncertainty in inverse problems.
We place our method in context with previous ideas. Our variational formulation shares
similarities with variational Bayes methods used in nonlinear inverse problems [Franck
and Koutsourelakis,2016] and stochastic design [Koutsourelakis,2016]. In these works,
an objective involving a complex posterior distribution is minimized under the constraint
that the approximating distribution is easy to sample from. In contrast, we use the
variational problem to define the distribution of interest. Second, when the likelihood is
intractable, approximate Bayesian computation (ABC) has been proposed as a viable method
of approximating intractable likelihoods [Lyne et al.,2015,Zeng et al.,2019]. However, these
procedures can be computationally costly and do not address model misspecification.
Finally, several methods have been proposed for solving stochastic inverse problems and
nonparametric probability measure estimation. Gradient-based optimization methods have
been used to solve stochastic inverse problems [Narayanan and Zabaras,2004,Borggaard and