Over recent years, the availability of water depth data increases rapidly as such data are consistently collected from
in-situ measurements at gauging stations as well as by remote sensing [
51
], such as satellite radar altimetry [
52
]. In-situ
data are continuous in time and sparse in space [
53
]. In contrast, remote sensing data, while measured at large time
intervals, can cover great spatial domains at high resolutions [
50
]. For example, the upcoming Surface Water and
Ocean Topography (SWOT) will enable direct measurements of water depth at a spatial resolution of 50 m for all rivers
wider than 100 m [
54
,
55
]. Such data could potentially be directly incorporated into ML models to obtain downscaled
solutions [
56
]. Remote sensing and gauged observation data, along with numerical solutions, provide the basis for the
development of ML-based methods.
The ML technique has been explored to alleviate the limitations in the river modeling from various aspects. For example,
ML models are employed as surrogates or emulators of hydrological models to perform rapid simulations for real-time
predictions and model calibrations [
57
]. ML models are also proposed to predict hydrograph [
57
], maximum water
levels [
58
], infiltration [
59
], as well as real-time inundation maps [
60
,
61
]. Despite differences in implementation, these
methods, such as regression trees, random forests and neural networks, all treat the ML model as a black box. Model
predictions are made based on a set of input variables, such as precipitation and streamflow. Additionally, training an
emulator is usually expensive as it requires abundant including observations or multiple realizations of high-resolution
numerical simulations. As a result, some alternative ML algorithms are proposed to replace computationally demanding
components of a numerical scheme with a computationally efficient ML model. For example, the convolutional neural
network (CNN) is used to solve the Euler equation for iterative pressure correction in the velocity update algorithm
[
31
]. Similarly, artificial neural networks are used to estimate the velocity in the inertial wave equation [
62
]. However,
these ML models are still computationally demanding for training and also they require accurate training data to avoid
bad numerical solutions.
Emerged recently are the state-of-the-art physics-informed ML models. Such models are trained by enforcing the
relevant physics. Data, mathematical models and domain knowledge are seamlessly integrated and implemented through
regression networks [
63
]. Physics-informed neural networks (PINN) solve partial differential equations (PDEs) using
feed-forward neural network architectures and respect the corresponding physical laws [
64
]. The unknown solutions of
governing equations are inferred from the initial and boundary data of the state variables. The ML models trained with
physical constraints require much less data than traditional ML emulators. This approach thus greatly alleviates the
limitations of ML in model training caused by data scarcity. The PINN algorithm can also be used in inverse problems to
estimate parameters of a PDE with observations of state variables provided [
65
,
66
], as well as in developing surrogate
models from sparse data [
67
,
68
]. Specialized network architectures have been developed for PINN to meet specific
physical laws [
69
], accelerate the training process [
70
] and improve the model accuracy [
71
]. As the result, many new
variants of PINN were introduced recently, such as multiphysics PINN [
72
], Parareal PINN [
73
,
74
], and DeepONet
[75]. For a comprehensive review of PINN, please refer to [63].
Due to the illustrated strengths, PINN has been explored extensively in solving forward, inverse and data assimilation
problems in science and engineering [
63
], such as subsurface transport in porous media [
72
] , acoustic wave propagation
[
76
], solid mechanics [
77
] and heat transfer [
78
]. In particular, PINN demonstrates tremendous success in simulating
flow dynamics [
79
], including high-speed aerodynamic flows [
80
], incompressible flow governed by Navier Stokes
equations [
68
,
81
], laminar flows at low Reynolds numbers [
82
], and even fluid-structure interaction [
83
]. Previous
applications of PINN to the hydraulic modeling [
84
], despite limited, show that it can outperform Artificial Neural
Network (ANN) in flood simulations [85].
The ML-based downscaling is essentially a data assimilation problem with the ML models merging measurements and
numerical model outputs into the ML training. A key feature of PINN is its simplicity for assimilating observations
[
72
,
66
]. The time-varying observations and/or spatial snapshots at various spatiotemporal scales can be readily
incorporated into the PINN training. Based on this feature, we propose a data assimilation approach based on the
standard PINN framework and its variant to address problems in Fourier spaces (ff-PINN) [
86
]. The proposed PINN
computes the subgrid solution of a coarse model output and assimilated observations without modifying the numerical
algorithms or refining the mesh resolution. By incorporating the dynamics presented in the measurements, PINN is able
to emulate the nonlinear interactions of storm surge, tide and discharge within tidal rivers. We do not intend to replace
the numerical solver by PINN. Instead, we aim to build the PINN method upon the infrastructure of existing large-scale
models to provide downscaled solutions at cells of interest.
The PINN-based method has a few additional advantages over local-scale numerical models. PINN is mesh-free,
making it an efficient and flexible tool for downscaling. In practice, the downscaled solution at regions of interest may
be obtained by configuring a local numerical model or an emulator at fine scales. Even though PINN itself may not be
as efficient as the local model in terms of running speed, PINN does not require discretization and thus saves efforts
of generating meshes for the numerical configuration. The meshless formalism enables flexible implementation in
various domains as the partial derivatives are evaluated using automatic differentiation in Tensorflow [
87
]. Moreover,
3