
arXiv:2210.03426v1 [cs.LG] 7 Oct 2022
1
Certified machine learning: Rigorous a posteriori
error bounds for PDE defined PINNs
Birgit Hillebrecht∗and Benjamin Unger†
Abstract—Prediction error quantification in machine learning
has been left out of most methodological investigations of neural
networks, for both purely data-driven and physics-informed
approaches. Beyond statistical investigations and generic results
on the approximation capabilities of neural networks, we present
a rigorous upper bound on the prediction error of physics-
informed neural networks. This bound can be calculated without
the knowledge of the true solution and only with a priori available
information about the characteristics of the underlying dynamical
system governed by a partial differential equation. We apply
this a posteriori error bound exemplarily to four problems: the
transport equation, the heat equation, the Navier-Stokes equation
and the Klein-Gordon equation.
Index Terms—Physics-informed neural network, machine
learning, certification, a posteriori error estimator, Navier-Stokes
I. INTRODUCTION
Physics-informed machine learning is applied to numerous
highly complex problems, such as turbulence and climate
modeling [7], [44], [45], model predictive control [3], [33],
and Hamiltonian system dynamics [16]. The systematic study
of physics-informed machine learning as a method, however,
remains open [30]. Part of this is the question about the quality
of the forecast: how close is the prediction of the physics-
informed neural network (PINN) [36] to the actual solution?
Beyond statistical evaluations, such as in [22], first steps
to answer this question were taken in [18] by establishing
rigorous error bounds for PINNs approximating the solution
of ordinary differential equations. We extend these results
by deriving guaranteed upper bounds on the prediction error
for problems modeled by linear partial differential equations
(PDEs). Our methodology is thereby applicable to, e.g., the
previously mentioned use cases of PINNs in computational
science and engineering. Similar to [18], the error bound can
be computed a posteriori without knowing the actual solution.
Since there are extensive studies on the convergence and ap-
proximation properties of (physics-informed) neural networks
[4], [11], [20], [39], it is clear from a theoretical point of view
that a suitable neural network (NN) can be found with the
desired accuracy. These a priori results and the associated error
estimates [8], [19], [39] are fundamentally different compared
B. Hillebrecht acknowledges funding from the International Max Planck
Research School for Intelligent Systems (IMPRS-IS). B. Unger acknowledges
funding from the DFG under Germany’s Excellence Strategy – EXC 2075
– 390740016 and The Ministry of Science, Research and the Arts Baden-
W¨urttemberg, 7542.2-9-47.10/36/2. Both authors are thankful for support by
the Stuttgart Center for Simulation Science (SimTech).
∗,†:Stuttgart Center for Simulation Science,University of Stuttgart, Stuttgart,
Germany, {birgit.hillebrecht,benjamin.unger}@simtech.uni-stuttgart.de ;
∗ORCID: 0000-0001-5361-0505 †ORCID: 0000-0003-4272-1079
to the computable a posteriori error bounds we present here.
Moreover, we like to stress that our analysis does not rely on
statistical approaches or the Bayesian perspective but provides
a rigorous error bound, which serves as a certificate for
the NN.
Close in topic is [39] wherein a so called ’a posterior’
error estimate is obtained based on norm inequalities. But,
as stated in their conclusion, the estimates are not directly
computable due to unknown constants. In distinction, the error
estimator presented here is computable which we demonstrate
with several academic examples. Other earlier works [5], [31]
also differ from the following work in fundamental aspects,
such as that the error bounds derived in them are either not
guaranteed to hold, are restricted to a certain problem type, or
require discretization steps.
At this point, we would like to emphasize that the results
presented in Section III are independent of PINNs as the
chosen methodology but can be applied to other surrogate
modeling techniques. Nevertheless, since PINNs use the norm
of the residual as a regularization term during the training
process, they appear as a natural candidate to apply our
residual-based error estimator.
Notation: We use bold notation for vectors and vector-
valued functions and italic letters for operators, especially
we use Ito denote the identity operator. The notations
˙
u=∂tu=∂u
∂t are used interchangeably to denote partial
derivatives w.r.t. time. Similarily, we shorten derivatives w.r.t.
spatial variables and use conventional notation for the diver-
gence ∇ · φ= div(φ), the gradient ∇u= grad(u), and the
Laplace operator ∆u= div(grad(u)) w.r.t. to spatial variables
only. For normed spaces (X, k·kX),(Y, k·kY), and operator
A:D(A)⊆X→Ywe define the induced operator norm
kAkX,Y := supx∈D(A),kxkX6=0
kAxkY
kxkX. In case, X⊆Yand
k·kY=k·kX, we shorten the notation to kAkX, and drop the
subscript whenever the norm is clear from the context. For the
spatial domain Ω⊆Rd, we denote the boundary by ∂Ωand
define the size of Ωas kΩk=RΩ1dx. The considerd temporal
domain is denoted by T= [0, tf)with tf∈R+. We use
the conventional notation for spaces of p-integrable functions
Lp(Ω), and Sobolev spaces Wk,p(Ω) and Hk(Ω) = Wk,2(Ω).
II. PROBLEM DESCRIPTION
In the following, we aim at solving the boundary and inital
value problem (BIVP): find u:T×Ω→Rnsuch that
∂tu=Auin T×Ω,
u=u0in {t= 0} × Ω,
Bu=ubin T×∂Ω.
(1)