Kalman-Bucy-Informed Neural Network for System Identification
Tobias Nagel and Marco F. Huber
Abstract— Identifying parameters in a system of nonlinear,
ordinary differential equations is vital for designing a robust
controller. However, if the system is stochastic in its nature or if
only noisy measurements are available, standard optimization
algorithms for system identification usually fail. We present a
new approach that combines the recent advances in physics-
informed neural networks and the well-known achievements of
Kalman filters in order to find parameters in a continuous-time
system with noisy measurements. In doing so, our approach
allows estimating the parameters together with the mean value
and covariance matrix of the system’s state vector. We show
that the method works for complex systems by identifying the
parameters of a double pendulum.
I. INTRODUCTION
Controlling a dynamical system in a safe manner re-
quires a model that describes the system properties precisely.
Ordinary differential equations (ODEs) are often used to
satisfy this requirement. Besides setting up the corresponding
equation operators, it is also inevitable to identify the real-
valued coefficients that define the characteristics of the
system. Estimating these parameters by using measurements
is termed as “inverse problem” and can be a difficult task,
depending on the system’s complexity. This work presents a
new method that is capable of identifying unknown param-
eters in a nonlinear ODE system, based on noisy measure-
ments by using an extended Kalman-Bucy filter (EKBF) in
a machine learning framework.
For linear systems, the subspace-based state space identi-
fication methods are well established. They aim at finding a
linear state space model by using a regularized least-squares
algorithm [7]. If the system comprises nonlinear behavior,
the most straightforward solution approaches for parame-
ter identification are standard minimization techniques like
gradient-based [9] or gradient-free [1] methods. For a system
with noisy measurements, the problem becomes even more
difficult and requires incorporating stochastic moments in
the optimization. Raue et. al. summarize their experiences
of fitting measurements of biological systems to their cor-
responding ODE system by maximizing a log-likelihood
function that comprises a normally distributed measurement
noise [13]. However, these methods require a numerical so-
lution of the ODE, repeatedly for each optimization iteration.
Besides being very time consuming, this approach often fails
Tobias Nagel and Marco F. Huber are with the Fraunhofer Institute for
Manufacturing Engineering and Automation IPA, Center for Cyber Cog-
nitive Intelligence (CCI), 70569 Stuttgart, Germany {tobias.nagel,
marco.huber}@ipa.fraunhofer.de
Marco F. Huber is with the Institute of Industrial Manufacturing
and Management IFF, University of Stuttgart, 70569 Stuttgart, Germany
marco.huber@ieee.org
because of the system’s nonlinearity, the noise influence or
an unstable behavior in the numerical solution [6].
A possibility to circumvent the problem with a machine
learning approach is described in [16] by using a neural
network to improve a Kalman filter system in order to obtain
a better state estimate. Though, this does not give us the
actual system parameter values but compensates for model
errors. In 2017, Raissi et. al. presented how physics-informed
neural networks (PINNs) can be trained by using modern
automatic differentiation frameworks [12]. The approach
utilizes deep neural networks to discover and solve nonlinear
differential equation systems. This is achieved by training a
neural network to represent an approximate solution to the
differential equation. The method also enables a parameter
search, by including the unknown parameters as additional
network weights. The concept has been applied in numerous
research fields, e.g., mechanics [8], thermodynamics [10] or
in chemical reaction equations [4]. PINNs also enable the
possibility to include stochastic behavior in the modeling
process. Recently, this has been addressed by O’Leary et. al.
who incorporate a mean value of the state and its covariance
matrix in the framework, leveraging it to a stochastic physics-
informed neural network (SPINN) [11]. The authors do so by
propagating the first two central moments of a state variable
through the known differential equation systems. Afterwards
a neural network is trained in order to match these estimated
central moments to measured ones. However, the authors
do not address the problem of identifying parameters in the
system. Another option is to use a Bayesian neural network
(BNN) in a PINN environment which allows an embedding
of uncertainty and, hence, the usage of stochastic differential
equations. Yang et. al. use a BNN [18] to include noisy data
into a partial differential equation problem in order to solve
as well as identify the system [18]. However, BNNs are often
not capable of achieving the same approximation accuracy as
standard neural networks and are significantly more difficult
to train.
In this paper, we present a new physics-informed machine
learning approach that we call Kalman-Bucy-informed neural
network (KBINN). A Kalman-Bucy filter incorporates two
ODEs that describe the temporal evolution of the mean
value and the covariance matrix of the system’s state. In
our method, we include two neural networks that are im-
plemented in a PINN framework in order to approximate
a solution to the Kalman-Bucy equations. This allows an
implicit identification of unknown system parameters by
incorporating them into the network training. The rest of the
paper is organized as follows: In Section II, we give a short
mathematical formulation of the problem. Section III intro-
arXiv:2210.03424v1 [eess.SY] 7 Oct 2022