1 Introduction
Estimation of many state space models with nonlinear and/or non-Gaussian measurement equations
is computationally challenging (Gribisch and Hartkopf, 2022; Chan, 2022; Cross et al., 2021). The
likelihood function of these models involves a high-dimensional integral with respect to the state
variables which cannot be solved analytically, and hence renders maximum likelihood estimation
to be infeasible. As an alternative, exact Bayesian estimation methods allow for the computation
of the posterior distribution of the model parameters. These methods either use particle filtering
(Chopin et al., 2020), or sample from the augmented posterior of the model parameters and the states
using analytical filtering (Carter and Kohn, 1994). Both approaches can become computationally
costly, especially with high-dimensional state vectors or when dependence between the states and
the parameters is strong (Quiroz et al., 2022).
Variational Bayes (VB) methods can provide a scalable alternative to exact Bayesian methods.
Instead of sampling exactly from the posterior, VB calibrates an approximation to the posterior
via the minimization of a divergence function. However, off-the-shelf variational methods for state
space models, such as mean-field variational approximations, are known to be poor (Wang and
Titterington, 2004). Gaussian VB methods as proposed by Tan and Nott (2018) and Quiroz et al.
(2022) use a variational family for the states that conditions on the model parameters and not on
the data. The inaccuracy of these existing methods is due to the quality of the approximation to
the conditional posterior distribution of the states (Frazier et al., 2022). More accurate VB methods
are computationally infeasible for many state space models. For instance, Tran et al. (2017) exactly
integrate out the states in the variational approximation using particle filtering. The method of
Loaiza-Maya et al. (2022) is designed for the specific class of state space models where generation
from the conditional posterior of the states is computationally feasible.
This paper proposes a novel VB method that is accurate and fast, and can be applied to a
wide range of state space models for which estimation with existing methods is either inaccurate or
computationally infeasible. Our method uses a variational approximation to the states that directly
conditions on the observed data, and as such produces an accurate approximation to the exact
posterior distribution. The approach is faster than existing VB methods for state space models
due to the computationally efficient calibration steps it entails. The implementation only requires a
measurement equation with a closed-form density representation, and a state transition distribution
that belongs to the class of exponential distributions. This allows for a wide range of state space
models, including ones with nonlinear measurement equations, certain types of nonlinear transition
equations, and high-dimensional state vectors.
Our approximation to the states is the importance density proposed by Richard and Zhang
(2007) in the context of efficient importance sampling. Hence, we refer to our method as Efficient
VB. Scharth and Kohn (2016) employ this importance distribution within a particle Markov chain
Monte Carlo (PMCMC) sampler to reduce the variance of the estimate of the likelihood function.
The use of this importance density inside PMCMC does not result in substantial computational
1