
Fourier Neural Solver for large sparse linear algebraic systems
et al. [
15
] exploited a CNN to design a more sensible smoother. Results showed that the magnitude of the learned
smoother is dispersed along the anisotropic direction. Wang et al. [
16
] introduced a learning-based local weighted least
square method for the AMG interpolation operator, and applied it to random diffusion equations and one-dimensional
small wavenumber Helmholtz equations. Fanaskov [
17
] learned the smoother and transfer operator of GMG in a
neural network form. When the anisotropic strength is mild (within two orders of magnitude), previously mentioned
works exhibit a considerable acceleration. Chen et al. [
18
] proposed the Meta-MgNet to learn a basis vector of Krylov
subspace as the smoother of GMG for strong anisotropic cases. However, the convergence rate is still sensitive to the
anisotropic strength. For convection-diffusion equations, Katrutsa et al. [
19
] learned the weighted Jacobi smoother and
transfer operator of GMG, which has a positive effect on the upwind discretazation system and also applied to solve a
one-dimensional Helmholtz equation. For second-order elliptic equations with random diffusion coefficients, Greenfeld
et al. [
20
] employed a residual network to construct the prolongation operator of AMG for uniform grids. Luz et al. [
21
]
extended it to non-uniform grids using graph neural networks, which outperforms classical AMG methods. For jumping
coefficient problems, Antonietti et al. [
22
] presented a neural network to forecast the strong connection parameter to
speed up AMG and used it as a preconditioner for CG. For Helmholtz equation, Stanziola et al. [
23
] constructed a
fully learnable neural solver, the helmnet, which is built on U-net and recurrent neural network [
24
]. Azulay et al. [
25
]
developed a preconditioner based on U-net and shift-Laplacian MG [
26
] and applied the flexible GMRES [
27
] to solve
the discrete system. For solid and fluid mechanics equations, there are also some neural solvers on associated discrete
systems, such as but not limited to, learning initial values [
28
,
29
], constructing preconditioners [
30
], learning search
directions of CG [31], learning parameters of GMG [32, 33].
In this paper, we propose the Fourier Neural Solver (FNS), a deep learning and Fast Fourier Transform (FFT) [
34
]
based neural solver. FNS is made up of two modules: the stationary method and the frequency space correction. Since
stationary methods like weighted Jacobi method are difficult to get rid of low-frequency error, FNS uses FFT and CNN
to learn these modes in the frequency space. Local Fourier analysis (LFA) [
5
] reveals that FNS can pick up on the
error components in frequency space that are challenging to eradicate by stationary methods. Therefore, FNS builds a
complementary relationship by stationary method and CNN to eliminate error. With the help of FFT, the single-step
iteration of FNS has a
O(Nlog2N)
computational complexity. All matrix-vector products are implemented using
convolution, which is both storage-efficient and straightforward to parallelize. We investigate the effectiveness and
robustness of FNS on three types of convection-diffusion-reaction equations. For anisotropic equations, numerical
experiments show that FNS can reduce the number of iterations by nearly
10
times compared to the state-of-the-art
Meta-MgNet when the anisotropic strength is relatively strong. For the non-symmetric systems arising from the
convection-diffusion equations discretized by central difference method, FNS can converge while MG and CG methods
diverge. And FNS is faster than other algorithms such GMRES and BiCGSTAB(
`
) [
35
]. For the indefinite systems
arising from the Helmholtz equations, FNS outperforms GMRES and BiCGSTAB at medium wavenumbers. In this
paper, we apply FNS to above three PDE systems. However, the principles employed by FNS indicate that FNS has the
potential to be useful for a broad range of sparse linear algebraic systems.
The rest of this paper is organized as follows. Section 2 proposes a general form of linear convection-diffusion-reaction
equation and describes the motivation for designing FNS. Section 3 presents the FNS algorithm. Section 4 examines
the performance of FNS to anisotropy, convection-diffusion, and Helmholtz equations. Finally, Section 5 draws the
conclusions and future work.
2 Motivation
We consider the general linear convection-diffusion-reaction equation with Dirichlet boundary condition
−ε∇ · (α(x)∇u) + ∇ · (β(x)u) + γu =f(x)in Ω
u(x) = g(x)on ∂Ω(2.1)
where
Ω⊆Rd
is an open and bounded domain.
α(x)
is the
d×d−
order diffusion coefficient matrix.
β(x)
is the
d×1velocity field that the quantity is moving with. γis the reaction coefficient. fis the source term.
We can obtain a linear algebraic system once we discretize Eq.
(4.1)
by finite element method (FEM) or finite difference
method (FDM)
Au =f,(2.2)
where A∈RN×N,f∈RNand Nis the spatial discrete degrees of freedom.
Classical stationary iterative methods, such as Gauss-Seidel and weighted Jacobi methods, have the generic form
uk+1 =uk+Bf−Auk,(2.3)
2