The Model Forest Ensemble Kalman Filter Andrey A. PopovyzandAdrian Sanduz Abstract. Traditional data assimilation uses information obtained from the propagation of one physics-driven

2025-05-06 0 0 2.12MB 26 页 10玖币
侵权投诉
The Model Forest Ensemble Kalman Filter
Andrey A. Popov,† ‡ and Adrian Sandu
Abstract. Traditional data assimilation uses information obtained from the propagation of one physics-driven
model and combines it with information derived from real-world observations in order to obtain a
better estimate of the truth of some natural process. However, in many situations multiple simulation
models that describe the same physical phenomenon are available. Such models can have different
sources. On one hand there are theory-guided models are constructed from first physical principles,
while on the other there are data-driven models that are constructed from snapshots of high fidelity
information. In this work we provide a possible way to make use of this collection of models in
data assimilation by generalizing the idea of model hierarchies into model forests—collections of
high fidelity and low fidelity models organized in a groping of model trees such as to capture various
relationships between different models. We generalize the multifidelity ensemble Kalman filter that
previously operated on model hierarchies into the model forest ensemble Kalman filter through a
generalized theory of linear control variates. This new filter allows for much more freedom when
treading the line between accuracy and speed. Numerical experiments with a high fidelity quasi-
geostrophic model and two of its low fidelity reduced order models validate the accuracy of our
approach.
Key words. Bayesian inference, control variates, data assimilation, multifidelity, ensemble Kalman filter, re-
duced order modeling
MSC codes. 62F15, 62M20, 65C05, 65M60, 76F70, 86A22, 93E11
1. Introduction. In many situations the availability of multiple models that describe the
same physical system is a valuable asset for obtaining accurate forecasts. For example the
Coupled Model Intercomparison Project [10] used by the International Panel on Climate
Change is an effort to utilize an aggregate of a wide array of climate models for the purposes
of increasingly accurate predictions. It is a recognition by the climate community that a
collection of models is greater than the sum of its parts.
The idea of leveraging a collection of models to improve data assimilation [1,20,28] has
seen an explosion of research over the last several years. Multilevel data assimilation was
first developed in the context of Monte Carlo methods [13,14], wherein a hierarchy of models,
through successive coarsening in the time dimension, was used to perform inference with
the accuracy of the finest level coarsening with a larger and larger amount of samples from
the coarser levels. The ideas of multilevel Monte Carlo were transferred to the ensemble
Kalman filter (EnKF) in a series of works developing the multilevel ensemble Kalman filter
(MLEnKF) [4,5,1719] aiming to provide more operationally viable methods.
Submitted to the ArXiv October 24, 2022.
Funding: The work of Popov and Sandu was supported by DOE through award ASCR DE-SC0021313, by NSF
through award CDS&E–MSS 1953113, and by the Computational Science Laboratory at Virginia Tech.
Oden Institute for Computational Engineering and Sciences, University of Texas at Austin, Austin, TX
(apopov@vt.edu)
Computational Science Laboratory, Department of Computer Science, Virginia Tech, Blacksburg, VA
(sandu@cs.vt.edu).
1
arXiv:2210.11971v1 [cs.CE] 21 Oct 2022
2 A. A. POPOV AND A. SANDU
The multifidelity ensemble Kalman filter (MFEnKF) [7,23,25,26] circumvents numerical
difficulties present in the MLEnKF through a robust use of linear control variate theory. The
MFEnKF also extends the idea of model coarseness to arbitrary non-linear couplings between
high fidelity (fine level) and low fidelity (coarse level) model states, allowing the use of various
types of reduced order models (ROMs) to form a model hierarchy.
This work further extends the EnKF ideas and brings two novel contributions. (i) First,
it extends model hierarchies to model trees and model forests, covering the situation were the
collection of models cannot neatly form a model hierarchy. (ii) Second, it extends the multifi-
delity ensemble Kalman filter to the model forest Kalman filter allowing data assimilation to
make use of model forests in a rigorous way.
Given one high fidelity model and a collection of low fidelity models, it is not always
possible to organize them in a strict model hierarchy. Following this observation we introduce
the first key contribution of the this work (i); we generalize the idea of model hierarchies to
model trees, where one model is allowed to have multiple low fidelity models on the same level
below it; the low fidelity models are surrogates for the high fidelity one, but they may not
have a direct relationship with each other. This results in a tree structure of models with the
high fidelity model acting as the root. We further extend model trees by leveraging the idea of
model averaging [8]. Assuming that we have a collection of model trees, each with their own
high fidelity model at the root, we organize them in a “model forest” and build an averaging
procedure over all the trees in the forest.
By bringing together the ideas of the MFEnKF with that of model forests, we make the
second key contribution (ii) of this work; we replace the MFEnKF with the model forest
ensemble Kalman filter, which also has the acronym MFEnKF as we show that the former is
a special case of the latter.
Numerical tests on the Quasi-Geostrophic equations with a quadratic reduced order model
and an autoencoder-based surrogate show that our proposed extension significantly decreases
the number of high fidelity model runs required to achieve a certain level of analysis accuracy.
This paper is organized as follows. Relevant background information including the se-
quential data-assimilation problem, model hierarchies, model averages, and the multifidelity
ensemble Kalman filter are presented in Section 2. The extension of model hierarchies to
model trees, and the extension of model averages to model forests is described in Section 3.
Next the extension of the multifideity ensemble Kalman filter to the model forest Kalman
filter is explained in Section 4. The quasi-geostrophic equations and two surrogate models are
detailed in Section 5. Numerical experiments on various model trees and model forests are
presented in Section 6. Finally, some closing remarks are stated in Section 7.
2. Background. We review relevant background on data assimilation, including model
hierarchies, linear control variates, model averaging, and the multifidelity ensemble Kalman
filter.
2.1. Data Assimilation. Let Xt
idenote the state of some natural process at time ti, where
the superscript t represents ground-truth. Assume that we have some prior information about
this state represented by the distribution of the random variable Xb
i. Assume also that we
MODEL FOREST ENKF 3
have access to some sparse noisy observations of the truth represented by,
(2.1) Yi=H(Xt
i) + εi,
where His a non-linear observation operator and iis a random variable representing obser-
vation error. For the remainder of this paper we assume that the observation error is normal
with distribution
(2.2) εi∼ N(0,ΣYi,Yi).
Finally, assume we have some inexact numerical model Mthat approximates the dynamics
of the natural process, i.e., evolution of the truth,
(2.3) Xt
i=M(Xt
i1) + ξi,
where the random variable ξirepresents the model error.
Data assimilation [1,9,28] seeks to combine the prior information Xb
iwith the sparse
noisy observations Yiinto a posterior representation Xa
iof the information, commonly through
Bayesian inference,
(2.4) π(Xa
i) = π(Xb
i|Yi)π(Yi|Xb
i)π(Xb
i),
where the distribution π(Xa
i) represents our full knowledge about the state of the system at
time ti.
The model (2.3) also forecasts the posterior information at time index iito prior
information at time i, through the relation,
(2.5) Xb
i=M(Xa
i1) + ξi.
2.2. Notation. In this work, the mean of the random variance Xis denoted by, µX, and
the covariance between the random variable Xand the random variable Yis denoted by, ΣX,Y .
An ensemble of Nsamples from the random variable Xis denoted by, EX= [X1,X2,...,XN],
with the ensemble mean denoted by,
e
µX=
N
X
i=1
1
NXi,
the scaled ensemble anomalies denoted by,
AX=1
N1EXe
µX1T
N,
where 1Nis a column vector of Nones, and the unbiased sample covariance between Xand
Ydenoted by, e
ΣX,Y =AXAT
Y.
4 A. A. POPOV AND A. SANDU
M(1)
M(1,1)
M(1,1,1)
Figure 2.1. A visual representation of a model hierarchy with two surrogate models. The principal model
,M(1), has a surrogate model , M(1,1), which in turn has its own surrogate model , M(1,1,1).
2.3. Model Hierarchies and Order Reduction. Assume there exists a model which is
expensive to compute from which we are attempting to glean some information through a
sampling procedure. Call this model the principal model. Assume that there exists a surrogate
model with which we can bootstrap our knowledge about the principal model. We can then
assume that the previously mentioned surrogate model is its own principal model in its own
model hierarchy that has its own surrogate model. This process can be repeated ad infinitum
to obtain a model hierarchy of a desired size. Figure 2.1 provides an illustration of a model
hierarchy for one principal model which has a surrogate that itself has a surrogate.
Let the tuple Irepresent the index of a model in the model hierarchy, such that the model
MIhas a surrogate model MI·1, with ·’ representing tuple concatenation, e.g., (1,2) ·3 =
(1,2,3). This particular notation helps with defining model trees and model forests later.
We make the following assumptions:
The dynamics of the high fidelity ‘principal’ model MIis embedded into the space
XI, i.e., MI:XIXI.
The dynamics of the low fidelity ‘surrogate’ model MI·1is embedded into the reduced
space XI·1, i.e., MI·1:XI·1XI·1.
There exists a (possibly non-linear) projection operator that maps the states of the
principal model to its surrogate:
(2.6) θI·1:XIXI·1.
There exists an interpolation operator that reconstructs an approximation of the state
of the principal model from that of the surrogate model:
(2.7) φI·1:XI·1XI,
The two operators obey the right-invertible consistency property [26],
(2.8) θI·1φI·1= id,
MODEL FOREST ENKF 5
ensuring that reconstruction has the same representation of the full order information
in the reduced space.
2.4. Linear Control Variates for Model Hierarchies. We discuss the specific case of a
bifidelity model hierarchy, , with the high fidelity model having one surrogate. Assume that
the information about our high fidelity model run is represented by the distribution of the
random variable Xknown as the principal variate. Assume also that there exist two random
variables whose distributions describe the information about the surrogate model: the control
variate b
Uwhich is highly correlated to X, and the ancillary variate Uwhich is uncorrelated
with the other variates, but shares its mean with b
U. The variates X,b
Uand Uare known as
the constituent variates.
Given some (possibly non-linear) functions hand g, the total variate which describes the
total information of the hierarchy in the linear control variate framework is given by,
(2.9) Zh=h(X)Shg(b
U)g(U)i
where Sis known as the gain operator. The choice of hand glargely depends on, and defines,
the information that is encapsulated by the different variates, and has to be carefully chosen
for each given problem.
Theorem 2.1. The optimal gain matrix Sthat minimizes the trace generalized variance of
Zin (2.9)is given by,
(2.10) S=Σh(X),g(b
U)Σg(b
U),g(b
U)+Σg(U),g(U)1.
Proof. By [22], the derivative with respect to Sof the trace generalized variance of Zis
Str(ΣZ,Z ) = 2Σh(X),g(b
U)+ 2SΣg(b
U),g(b
U)+Σg(U),g(U),
and as the Hessian is always symmetric positive definite,
2
S2tr(ΣZ,Z )=2Σg(b
U),g(b
U)+Σg(U),g(U)I0,
the global minimum is attained when,
(2.11)
Str(ΣZ,Z ) = 0,
which is satisfied by (2.10), as required.
We now describe the generalization to a model hierarchy. Assume that the ancillary variate
with indexing tuple Iis the total variate estimator for
(2.12) UI=hIXISI·1hgI·1b
UI·1gI·1UI·1i,
with Zh:=U(1) representing the total variate
摘要:

TheModelForestEnsembleKalmanFilterAndreyA.Popov,yzandAdrianSanduzAbstract.Traditionaldataassimilationusesinformationobtainedfromthepropagationofonephysics-drivenmodelandcombinesitwithinformationderivedfromreal-worldobservationsinordertoobtainabetterestimateofthetruthofsomenaturalprocess.However,inm...

展开>> 收起<<
The Model Forest Ensemble Kalman Filter Andrey A. PopovyzandAdrian Sanduz Abstract. Traditional data assimilation uses information obtained from the propagation of one physics-driven.pdf

共26页,预览5页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!

相关推荐

分类:图书资源 价格:10玖币 属性:26 页 大小:2.12MB 格式:PDF 时间:2025-05-06

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 26
客服
关注