In this paper, we propose to learn the continuous solutions for spatio-temporal PDEs. Previous
methods focused on either generating fixed resolution predictions or generating arbitrary resolution
solutions on a fixed grid (Li et al., 2021; Wang et al., 2020). PDE models based on Multi-Layer
Perceptrons (MLPs) can generate solutions at any point of the domain (Dissanayake and Phan-Thien,
1994; Lagaris et al., 1998; Raissi et al., 2017a). However, without imposing a physics-motivated loss
that constrains the predictions to follow the smoothness bias resulting from the PDE, MLPs become
less competitive than CNN-based approaches especially when the PDE solutions have high-frequency
information (Rahaman et al., 2018).
We leverage the recent advances in Implicit Neural Representations ((Tancik et al., 2020), (Chen et al.,
2020), (Jiang et al., 2020)) and propose a general purpose model that can not only learn solutions to a
PDE with a resolution it was trained on, but it can also perform zero-shot super-resolution on irregular
meshes. The added advantage is that we propose a general framework where we can make predictions
given any spatial position query for both grid-based architectures like CNNs and graph-based ones
able to handle sensors and predictions at arbitrary spatial positions.
Contributions
Our main contributions are in the context of machine learning for approximately
but efficiently solving PDEs and can be summarized as follows:
•
We propose a framework that enables grid-based and graph-based architectures to generate
continuous-space PDE solutions given a spatial query at any position.
•
We show experimentally that this approach can generalize to resolutions up to four times
those seen during training in zero-shot super-resolution tasks.
2 Related Works
Current solvers can require a lot of computations to generate solutions on a fine spatio-temporal
grid. For example, climate predictions typically use General Circulation Models (GCM) to make
forecasts that span several decades over the whole planet (Phillips, 1956). These GCMs use PDEs to
model the climate in the atmosphere-ocean-land system and to solve these PDEs, classical numerical
solvers are used. However, the quality of predictions is bottlenecked by the grid resolution that is in
turn constrained by the available amount of computing power. Deep learning has recently emerged
as an alternative to these classical solvers in hopes of generating data-driven predictions faster and
making approximations that do not just rely on lower resolution grids but also on the statistical
regularities that underlie the family of PDEs being considered. Using deep learning also makes it
possible to combine the information in actual sensor data with the physical assumptions embedded in
the classical PDEs. All of this would enable practitioners to increase the actual resolution further for
the same computational budget, which in turn improves the quality of the predictions.
Machine Learning for PDE solving:
Dissanayake and Phan-Thien (1994) published one of the
first papers on PDE solving using neural networks. They parameterized the solutions to the Poisson
and heat transfer equations using an MLP and studied the evolution of the error with the mesh size.
Lagaris et al. (1998) used MLPs for solving PDEs and ordinary differential equations. They wrote
the solution as a sum of two components where the first term satisfies boundary conditions and is
not learnable, and the second is parameterized with an MLP and trained to satisfy the equations. In
Raissi et al. (2017a) the authors also parameterized the solution to a PDE using an MLP that takes
coordinates as input. With the help of automatic differentiation, they calculate the PDE residual and
use its MSE loss along with an MSE loss on the boundary conditions. In follow-up work, Raissi et al.
(2017b) also learn the parameters of the PDE (e.g. Reynolds number for Navier-Stokes equations).
The recently introduced Neural Operators framework (Kovachki et al., 2021; Li et al., 2020b,a)
attempts to learn operators between spaces of functions. Li et al. (2021) use "Fourier Layers" to
learn the solution to a PDE by framing the problem as learning an operator from the space of initial
conditions to the space of the PDE solutions. Their model can learn the solution to PDEs that lie on a
uniform grid while maintaining their performance in the zero-shot super-resolution setting. In the
same spirit, Jiang et al. (2020) developed a model based on Implicit Neural Representations called
"MeshFreeFlowNet" where they upsample existing PDE solutions to a higher resolution. They use
3D low-resolution space-time tensors as inputs to a 3DUnet in order to generate a feature map. Next,
some points are sampled uniformly from the corresponding high-resolution tensors and fed to an
MLP called ImNet (Chen and Zhang, 2018). They train their model using a PDE residual loss and
2