
2
Brehmer and Louppe,2020). When a proper prior can be
specified, we may simulate parameters and random com-
ponents independently, obtain imputed data through the
DGE, and retain the samples if and only if the imputed
and observed data are sufficiently close. Such an accept-
reject scheme is often referred to as approximate Bayesian
computation (ABC; e.g., Beaumont,2019;Beaumont,
Zhang and Balding,2002;Beaumont et al.,2009;Blum,
2010;Fearnhead and Prangle,2012;Marin et al.,2012;
Sisson and Fan,2011;Sisson, Fan and Beaumont,2018):
The retained samples of parameters approximately follow
the posterior distribution and hence can be utilized to es-
timate posterior expectations. If no prior distribution is
available, we can still sample random components but not
parameters. To circumvent the latter, GFI proceeds to pair
each realization of random components with the optimal
parameter values such that the resulting imputed data is as
close to the observed data as possible in some sense. In-
deed such a best matching to the observed data may still
not be good enough: Those values are deemed incompat-
ible with the observed data and therefore have to be dis-
carded, leading to a rejection step similar to ABC. It turns
out that the resulting marginal samples of parameters ap-
proximately follow the GFD (Hannig et al.,2016).
It is then natural to ponder what the limits of the trun-
cated distributions are when we request the imputed data
to be infinitesimally close to the observed data in ap-
proximate post-data inference. As the main result of the
present work, we completely characterize the weak limit
for both approximate Bayesian inference and GFI when
the truncation set contracts to a twice continuously dif-
ferentiable submanifold of the joint space of parameters
and random components. We are able to express the ab-
solutely continuous densities of the limiting distributions
with respect to the intrinsic measure of the submanifold,
and show that Bayesian posteriors and GFDs in the usual
sense are the corresponding marginals on the parameter
space (Propositions 1and 2). As a contribution to the lit-
erature of GFI, we derive an explicit formula for the fidu-
cial density in Proposition 2that is more general com-
pared to Theorem 1 of Hannig et al. (2016). Meanwhile,
our work should be distinguished from Murph, Hannig
and Williams (2022b), which also studied the geometry
of GFI but focused on the case when the parameter space
itself is a manifold. On the theoretical side, our geomet-
ric formulation applies to a broad class of parametric sta-
tistical models for continuous data and facilitates insight-
ful comparisons between Bayesian inference and GFI. On
the practical side, the geometric characterization suggests
an alternative sampling scheme for approximate post-data
inference: We apply manifold Markov chain Monte Carlo
generative process rather than the formal mathematical expression is
of interest.
(MCMC) algorithms (e.g., Brubaker, Salzmann and Urta-
sun,2012;Zappa, Holmes-Cerfon and Goodman,2018)
to sample from the limiting distributions on the data gen-
erating manifold and only retain the parameter marginals.
For certain problems (e.g., GFI for mixed-effects mod-
els), manifold MCMC sampling may scale up better than
existing computational procedures.
The rest of the paper is organized as follows. We re-
visit in Section 2the formal definitions of ABC and GFI;
a graphical illustration is provided using a Gaussian lo-
cation example. In Section 3, we first present a general
result (Theorem 1): When an ambient distribution is trun-
cated to a sequence of increasingly finer approximations
to a smooth manifold, the weak limit is absolutely con-
tinuous with respect to the manifold’s intrinsic measure.
We then apply the general result to derive representations
for Bayesian posteriors and GFDs (Propositions 1and 2)
and comment on their discrepancies. We review in Sec-
tion 4an MCMC algorithm that (approximately) samples
from distributions on differentiable manifolds. A repeated
measures analysis of variance (ANOVA) example is then
presented to illustrate the sampling procedure (Section
5). Limitations and possible extensions of the proposed
method are discussed at the end (Section 6).
2. APPROXIMATE INFERENCE BY SIMULATION
2.1 Data Generating Equation
Let Y,Υ, and Θdenote the spaces of data, random
components, and parameters associated with a fixed fam-
ily of parametric models: In particular, Y ⊆Rn,Υ⊆Rm,
and Θ⊆Rq, where n,m, and qare positive integers. Fol-
lowing Hannig et al. (2016), we characterize the model of
interest by its DGE
(1) Y=G(U, θ),
in which the random components U∈Υfollow a com-
pletely known distribution (typically uniform or standard
Gaussian), θ∈Θdenotes the parameters, and Y∈ Y de-
notes the random data. (1) can be conceived as a formal-
ization of the data generating code: Given true parameters
θand an instance of random components U=u, a unique
set of data Y=ycan be imputed by evaluating the DGE,
i.e., y=G(u, θ).
Now suppose that we have observed Y=y. Post-data
inference aims to assign probabilities to assertions about
parameters θconditional on the observed data y(Mar-
tin and Liu,2015c). In the conventional Bayesian frame-
work, we presume that θfollows a proper prior distribu-
tion and make probabilistic statements based on the con-
ditional distribution of θgiven y. When it is difficult to
specify an informative prior, one may still rely on objec-
tive priors that reflect paucity of knowledge or informa-
tion (Kass and Wasserman,1996;Berger,2006;Berger,