
1 Introduction
In this paper we demonstrate the applicability of the recently proposed nonlinear sparse Bayesian learning (NSBL)
algorithm [
1
,
2
] to a single degree of freedom (SDOF) aeroelastic oscillator that is undergoing low amplitude limit
cycle oscillations. This experimental setup has been studied extensively through experimentation [
3
,
4
], numerical
modelling of laminar separation flutter using high-fidelity large eddy simulations (LES) [
5
] and unsteady Reynolds
averaged Navier-Stokes (URANS) model [
6
]. Furthermore, the wind tunnel experiments have provided a reliable
test-bed for developing Bayesian techniques for system identification and model selection for nonlinear dynamical
systems [
7
,
8
,
9
,
10
]. The work in [
7
,
8
,
9
] use standard methods of evidence-based Bayesian model selection, which
allows for the systematic comparison of a set of candidate models with varying degrees of complexity. The model
evidence as a criterion for model selection ensures a balance of favouring models with superior average data-fit, while
penalizing models that are overly complex and thus prone to overfitting [
11
]. In this context, model complexity is
quantified by the KL-divergence of the parameter posterior probability density function (pdf) from the parameter prior
pdf. For parameters where there exists little prior knowledge, it is typical to assign non-informative priors, however the
width of the distribution used for the non-informative prior will influence the optimal complexity of the model. The
issue of sensitivity to prior width is addressed in [
10
], whereby the problem is reposed as a sparse learning problem.
Rather than non-informative priors, parameters with little prior information are assigned Gaussian automatic relevance
determination (ARD) priors. The precision (inverse of the variance) of these ARD priors are determined through
evidence optimization. In this re-framing of the inference problem, the optimal model is still quantified as such based
on the model evidence. In contrast to the previous approach, rather than proposing an entire set of nested candidate
models to determine the optimal model complexity, the current paper approaches the problem as an automatic discovery
of the optimal sparse model nested within a single (potentially) over-parameterized model. Herein lies an additional
benefit of approaching the problem as a sparse learning task; it is only necessary to obtain the parameter posterior for a
single model, whereas standard methods require the calibration of each model in the candidate in order to then obtain an
estimate of the model evidence. The shortcoming of this approach lies in the fact that the optimization process involves
the use of Markov Chain Monte Carlo (MCMC) sampling at each iteration. This is addressed in the current NSBL
framework, which removes the use of MCMC from within the optimization loop, resulting in significantly improved
computational efficiency.
The NSBL framework presented here is an extension of the sparse Bayesian learning (SBL) also known as the relevance
vector machine (RVM) algorithm [
12
,
13
]. Both methods are motivated by the desire to avoid overfitting during
Bayesian inversion. SBL/RVM and the similar Bayesian compressive sensing (BCS) algorithm [
14
] provide analytical
expressions for a sparse parameter posterior distribution owing to the analytical conveniences of the semi-conjugacy
that exists between the Gaussian likelihood functions, and Gaussian ARD priors that are conditioned on hyperpriors
that are Gamma distributions. The SBL methodology is extended to be applicable to nonlinear-in-parameter models
and for non-Gaussian prior distributions, as these both commonly arise in engineering applications. We provide the
minimum required mathematical details to understand the objectives of the algorithm and to provide a complete account
of all terms shown in the equations used in this paper. For the full detailed derivation and additional details, we refer the
reader to [1, 2].
2 Methodology: Nonlinear sparse Bayesian learning
The NSBL methodology is applicable to general nonlinear mappings,
f:φ7→ y
where the model operator
f
maps the
unknown model parameter vector
φ∈RNφ
to the observable model output
y∈RNy
. In this specific application,
f
represents the aeroelastic model,
φ
are the deterministic system parameters and the stochastic parameters (relating to
the model error), and
y
are the system output. Sensor measurements of the system output
y
at discrete points in time
are denoted as
D
. The likelihood function
p(D|φ)
can be computed for any
φ
, using the observations
D
, and these
observation may be noisy, sparse, and incomplete measurements of the system state. The purpose of the algorithm is to
obtain a data-optimal sparse representation of φusing Bayesian inversion, while removing redundant parameters.
NSBL operates within the following Bayesian framework; we seek the posterior distribution of the unknown model
parameters φconditioned on the data Dand hyperparameters α,
p(φ|D,α) = p(D|φ)p(φ|α)
p(D|α)=p(D|φ)p(φ|α)
Rp(D|φ)p(φ|α)dφ(1)
for given data and hyperparameters, the denominator, which represents the model evidence (or marginal likelihood or
type-II likelihood), is just a normalization constant. The parameter prior
p(φ|α)
is also conditional on the hyperparam-
eter. Though the objective is not to perform full hierarchical Bayesian inference, we nevertheless define a prior on
p(α)
(which is notably absent in the expression above); this hyperparameter prior (or hyperprior) becomes relevant during
the optimization of α.
2