Bayesian probability updates using SamplingImportance Resampling Applications in nuclear theory W. G. Jiang and C. Forss en

2025-04-27 0 0 7.82MB 19 页 10玖币
侵权投诉
Bayesian probability updates using Sampling/Importance
Resampling: Applications in nuclear theory
W. G. Jiang and C. Forss´en
Department of Physics, Chalmers University of Technology, SE-412 96 G¨oteborg, Sweden
Abstract
We review an established Bayesian sampling method called sampling/importance resampling
and highlight situations in nuclear theory when it can be particularly useful. To this end we both
analyse a toy problem and demonstrate realistic applications of importance resampling to infer the
posterior distribution for parameters of ∆NNLO interaction model based on chiral effective field
theory and to estimate the posterior probability distribution of target observables. The limitation
of the method is also showcased in extreme situations where importance resampling breaks.
PACS numbers: 21.30.-x
1
arXiv:2210.02507v1 [nucl-th] 5 Oct 2022
I. INTRODUCTION
Bayesian inference is an appealing approach for dealing with theoretical uncertainties
and has been applied in different nuclear physics studies [
1
14
]. In the practice of Bayesian
analyses, a sampling procedure is usually inevitable for approximating the posterior probability
distribution of model parameters and for performing predictive computations. Various
Markov chain Monte Carlo (
MCMC
) methods [
15
19
] are often used for this purpose, even
for complicated models with high-dimensional parameter spaces. However,
MCMC
sampling
typically requires many likelihood evaluations, which is often a costly operation in nuclear
theory, and there is a need to explore other sampling techniques. In this paper, we review an
established method called sampling/importance resampling (
S/IR
) [
20
22
] and demonstrate
its use in realistic nuclear physics applications where we also perform comparisons with
MCMC sampling.
In recent years, there has been an increasing demand for precision nuclear theory.This
implies a challenge to not just achieve accurate theoretical predictions but also to quantify
accompanying uncertainties. The use of ab initio many-body methods and nuclear interaction
models based on chiral effective field theory (
χ
EFT) has shown a potential to describe
finite nuclei and nuclear matter based on extant experimental data (e.g. nucleon-nucleon
scattering, few-body sector) with controlled approximations [
23
27
]. The interaction model is
parametrized in terms of low-energy constants (
LEC
s), the number of which is growing order-
by-order according to the rules of a corresponding power counting [
28
30
]. Very importantly,
the systematic expansion allows to quantify the truncation error and to incorporate this
knowledge in the analysis [
4
6
,
10
14
]. Indeed, Bayesian inference is an excellent framework
to incorporate different sources of uncertainty and to propagate error bars to the model
predictions. Starting from Bayes’ theorem
pr(θ|D)∝ L(θ)pr(θ),(1)
where
pr
(
θ|D
) is the posterior probability density function (
PDF
) for the vector
θ
of
LEC
s
(conditional on the data
D
),
L
(
θ
)
pr
(
D |θ
) is the likelihood and
pr
(
θ
) is the prior. Then
for any model prediction one needs to evaluate the expectation value of a function of interest
y(θ) (target observables) according to the posterior. This involves integrals such as
Zdθy(θ)pr(θ|D),(2)
2
which can not be analytically solved for realistic cases. Fortunately, integrals such as Eq.
(2)
can be approximately evaluated using a finite set of samples
{θi}N
i=1
from
pr
(
θ|D
).
MCMC
sampling methods are the main computational tool for providing such samples, even for
high-dimensional parameter volumes [
31
]. However the use of
MCMC
in nuclear theory
typically requires massive computations to record sufficiently many samples from the Markov
chain. There are certainly situations where
MCMC
sampling is not ideal, or even becomes
infeasible:
1.
When the posterior is conditioned on some calibration data for which our model
evaluations are very costly. Then we might only afford a limited number of full
likelihood evaluations and our MCMC sampling becomes less likely to converge.
2.
Bayesian posterior updates in which calibration data is added in several different stages.
This typically requires that the
MCMC
sampling must be carried out repeatedly from
scratch.
3.
Model checking where we want to explore the sensitivity to prior assignments. This is
a second example of posterior updating.
4.
The prediction of target observables for which our model evaluations become very
costly and the handling of a large number of MCMC samples becomes infeasible.
These are situations where one might want to use the
S/IR
method [
21
,
22
], which allows
posterior probability updates with a minimum amount of computation (previous results of
model evaluations remain useful). In the following sections we first review the
S/IR
method
and then present both toy and realistic applications in which its performance is compared
with full
MCMC
sampling. Finally, we illustrate limitations of the method by considering
cases where
S/IR
fails and we highlight the importance of the so-called effective number of
samples. More difficult scenarios, in which the method fails without a clear warning, are left
for the concluding remarks.
II. SAMPLING/IMPORTANCE RESAMPLING
The basic idea of
S/IR
is to utilize the inherent duality between samples and the density
(probability distribution) from which they were generated [
21
]. This duality offers an
3
opportunity to indirectly recreate a density (that might be hard to compute) from samples
that are easy to obtain. Here we give a brief review of the method and illustrate with a toy
problem.
Let us consider a target density
h
(
θ
). In our applications this target will be the posterior
PDF pr
(
θ|D
) from Eq.
(1)
. Instead of attempting to directly collect samples from
h
(
θ
), as
would be the goal in
MCMC
approaches, the
S/IR
method uses a detour. We first obtain
samples from a simple (even analytic) density
g
(
θ
). We then resample from this finite set
using a resampling algorithm to approximately recreate samples from the target density
h
(
θ
). There are (at least) two different resampling methods. In this paper we only focus on
one of them called weighted bootstrap (more details of resampling methods can be found in
Refs. [20, 21]).
Assuming we are interested in the target density
h
(
θ
) =
f
(
θ
)
/Rf
(
θ
) d
θ
, the procedure
of resampling via weighted bootstrap can be summarized as follows:
1. Generate the set {θi}n
i=1 of samples from a sampling density g(θ).
2.
Calculate
ωi
=
f
(
θi
)
/ g
(
θi
) for the
n
samples and define importance weights as:
qi=ωi/Pn
j=1 ωj.
3.
Draw
N
new samples
{θ
i}N
i=1
from the discrete distribution
{θi}n
i=1
with probability
mass qion θi.
4.
The set of samples
{θ
i}N
i=1
will then be approximately distributed according to the
target density h(θ).
Intuitively, the distribution of
θ
should be good approximation of
h
(
θ
) when
n
is large
enough. Here we justify this claim via the cumulative distribution function of
θ
(for the
one-dimensional case)
pr(θa) =
n
X
i=1
qi·H(aθi) =
1
n
n
P
i=1
ωi·H(aθi)
1
n
n
P
i=1
ωi
n→∞
Eghf(θ)
g(θ)·H(aθi)i
Eghf(θ)
g(θ)i=Ra
−∞ f(θ)
R
−∞ f(θ)=Za
−∞
h(θ),
(3)
4
摘要:

BayesianprobabilityupdatesusingSampling/ImportanceResampling:ApplicationsinnucleartheoryW.G.JiangandC.ForssenDepartmentofPhysics,ChalmersUniversityofTechnology,SE-41296Goteborg,SwedenAbstractWereviewanestablishedBayesiansamplingmethodcalledsampling/importanceresamplingandhighlightsituationsinnucle...

展开>> 收起<<
Bayesian probability updates using SamplingImportance Resampling Applications in nuclear theory W. G. Jiang and C. Forss en.pdf

共19页,预览4页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:19 页 大小:7.82MB 格式:PDF 时间:2025-04-27

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 19
客服
关注