
Autocorrelated measurement processes and inference for ODEs in biology A PREPRINT
There is a huge literature devoted to accounting for model misspecification during inference (see, for example, [13
–
16]),
and this remains an active area of research. In this paper, however, we focus only on the impact of assumptions around
measurement noise, since, as we demonstrate, these can have dramatic effects on inference even in the absence of model
misspecification. To exemplify how measurement process imperfections can lead to autocorrelation, suppose again that
model A is the true model of nature, and that we (correctly) use it as part of our model of the data generating process.
Also, suppose that the measuring apparatus is imperfect, producing noisy observations that may differ from the true
underlying state, and has finite temporal resolution meaning it struggles to capture changes in output over shorter time
scales. In Figs. 1B&C, we show the model solutions (solid lines) and the values that would be measured if using a very
fine temporal gridding (dashed lines). A consequence of this smooth measurement process is that the more observations
per unit time are taken, the greater the degree of autocorrelation in residuals. In Fig. 1B, we show coarse observations
of the system of interest as indicated by the horizontal positioning of the vertical arrows. In this case, since observations
are sufficiently separated in time, there is relatively low persistence in residuals. In Fig. 1C, we take more observations
of the same process, which produces positively autocorrelated residuals.
Intuitively, when the measurement process is positively autocorrelated, each observation conveys less information
about the system than when the observations are uncorrelated. So misrepresenting an autocorrelated error process
with one assuming independence can lead to overly confident parameter estimates. This is a well-known result in
regression modelling [17], and, since fitting ODE models to data is just nonlinear regression, these results should also
apply to inference for these model types. We show this in the inset panels in Figs. 1B&C: here, the orange lines show
(illustrative) posterior distributions resultant from modelling the measurement process correctly; the green lines show
the distributions when modelling the measurements assuming independence amongst them. In Fig. 1B, where the
measurements are widely spaced, there is little difference in the recovered posteriors due to the limited autocorrelation.
In Fig. 1C, failure to account for autocorrelation results in a posterior with too little variance.
We originally became interested in the impact of measurement autocorrelation on parameter estimation when attempting
inference for a model of an electrochemistry experiment. Specifically, we noticed that the estimates obtained were
unrealistically precise when assuming an IID normal error model, and the errors were autocorrelated. This led us to
consider how this phenomena might be more generally applicable and whether there were guiding principles of how the
degree of overconfidence depends on measurement autocorrelation. Thus, in this paper, we explore how measurement
autocorrelation affects the precision of estimates. Previous work, in the context of modelling physical systems, has
derived straightforward expressions for parameter uncertainty for a dataset consisting only of two time points with an
accordingly simple error model [12]. Here, we consider a much more general setting where the models are nonlinear
ODEs, which is typical in biological systems analysis, and the measurement process can be any one of a wide class of
stochastic processes. We also account for the bias in the estimates of the standard deviation of the noise when fitting a
model assuming IID Gaussian errors, which is important to ensure correct estimates of the degree of overconfidence.
Using simulated data from ODE models, we demonstrate the validity of our analytical results. Using experimental data
from cardiac physiology and electrochemistry, we show that highly persistent differences between observations and
modelled quantities can occur. Whilst only illustrative, these results hint that overconfidence in parameter estimates
may not be uncommon. In addition, we provide a workflow for diagnosing and accounting for autocorrelated errors
when fitting an ODE model to data.
3 Effect of autocorrelated noise on parameter estimate uncertainty
In this section, we use mathematical analysis to evaluate the effect on parameter estimates of not accounting for
autocorrelation when present. To do so, we first calculate “true” parameter uncertainties obtained when specifying
a persistent error model. We then calculate “false” uncertainties obtained when assuming independent errors. To
derive these quantities, we calculate the Fisher Information matrix (FIM) in both circumstances. This analysis shows
that uncertainty in parameter estimates is understated when (falsely) assuming independent errors, with the degree of
overconfidence increasing along with the persistence of the true errors. We call the ratio of true parameter estimate
variance to that estimated assuming independent errors the “variance inflation ratio” (VIR).
In
§
3.1, we estimate the VIR for the mean parameter of a simple model with constant mean, when the actual error
process is persistent and described by an autoregressive order-one (AR(1)) process. Calculating the VIR for the constant
mean model is straightforward but provides a useful guide when examining more realistic cases. In
§
3.2, we consider
a nonlinear ODE model with AR(1) measurement noise. In
§
3.3, we explore the consequences of more ephemeral
autocorrelations by calculating the VIR for the constant mean model with moving average order-one (MA(1)) errors.
Realistic noise processes are likely, in fact, to be combinations of persistent and transient correlated noise, and in
§
3.4,
we give formulae for computation of VIRs in this, more general, case.
3