manner, the ANN can be trained to learn data from a physical law that is given by a PDE or a system
of PDEs. The idea is quite similar to the classical Galerkin methods, but instead of representing the
solution as a projection in some flavour of Galerkin space, the solution is written in terms of ANNs
as the composition of nonlinear functions depending on some network weights. As a result, instead
of a high dimensional linear system, a high dimensional nonlinear optimization problem is obtained
for the ANN weights. This problem must be solved using nonlinear optimization algorithms such as
stochastic gradient descent-based methods, e.g., [45], and/or quasi-Newton methods, e.g., L-BFGS,
[7]. More recently, with the advances in automatic differentiation algorithms (AD) and hardware
(GPUs), this kind of techniques have gained more momentum in the literature and, currently, the
most promising approach is known as physics-informed neural networks (PINNs), see [53] [59], [51],
[54], [23].
In the last few years, PINNs have shown a remarkable performance. However, there is still some
room for improvements within the methodology. One of the disadvantages of PINNs is the lack of
theoretical results for the control of the approximation error. Obtaining error estimates or results for
the order of approximation in PINNs is a non-trivial task, much more challenging than in classical
methods. Even so, the authors in [25], [4], [54], [28], [26], [24] and [27] (among others) have derived
estimates and bounds for the so-called generalization error considering particular models. Another
drawback is the difficulty when imposing the boundary conditions (a fact discussed further later in
this section). Nevertheless the use of ANNs has several advantages for solving PDEs: they can be used
for nonlinear PDEs without any extra effort; they can be extended to (moderate) high dimensions;
and they yield accurate approximations of the partial derivatives of the solution thanks to the AD
modules provided by modern deep learning frameworks.
PINNs is not the only approach relying on ANNs to address PDE-based problems. They can be
used as a complement for classical numerical methods, for example training the neural network to
obtain smoothness indicators, or WENO reconstructions in order for them to be used inside a classical
FV method, see [46], [47]. Also ANNs are being used to solve PDE models by means of their backward
stochastic differential equation (BSDE) representation as long as the Feynmann-K`ac theorem can be
applied, which is the usual situation in computational finance, for example. In [37], the authors present
the so called DeepBSDE numerical methods and their application to the solution of the nonlinear
Black-Scholes equation, the Hamilton-Jacobi-Bellman equation, and the Allen-Cahn equation in very
high (hundreds of) dimensions. The connection of such method with the recursive multilevel Picard
approximations allows the authors to prove that DeepBSDEs are capable of overcoming the so called
“curse of dimensionality” for a certain kind of PDEs, see [68], [42].
The main goal of the present work is to develop robust and stable deep learning numerical methods
for solving nonlinear parabolic PDE models by means of PINNs. The motivation arises from the
difficulty of finding and numerically imposing the boundary conditions, which are always delicate
and critical tasks both in the classical FD/FV/FE setting and thus also in the ANN setting. The
common approach consists in assigning weights to the different terms involved in the loss function,
where the selection of this weights must be done heuristically. We introduce a new idea that consists
in introducing the loss terms due to the boundary conditions by means of evaluating the PDE operator
restricted to the boundaries. In this way the value of such addends is of the same magnitude of the
interior losses. Although this is non feasible in the classical PDE solving algorithms, it is very intuitive
within the PINNs framework since, by means of AD, we can evaluate this operator in the boundary
even in the case it contains normal derivatives to such boundary. Thus, this novel treatment of the
boundary conditions in PINNs is the main contribution of this work, allowing to get rid of the heuristic
choice of the weights for the contributions of the boundary addends to the loss function that come from
the boundary conditions. Further, AD can be naturally exploited to obtain accurate approximations of
the partial derivatives of the solution with respect to the input parameters (quantities of much interest
in several fields).
Although the proposed methodology could be presented for a wide range of applications, here we
will focus on the solution of PDE models for challenging problems appearing the the computational
finance field. In particular, we consider the derivative valuation problem in the presence of counterparty
2