
learning method is demanded to replace the high-cost operators of graph convolutions and attention
(with quadratic time complexity with the graph size); 3) the graph over MTS varies with different
temporal interactions, which demands an efficient dynamic structure encoding method. In view of
the increasing leverage of the convolution theorem in Fourier theory to design high-efficiency neural
units (operators) to replace costly attention computations Guibas et al. (2022), we are inspired to
perform graph computations in the Fourier space and provide a potential alternative for processing
supra-graphs jointly attending to spatial-temporal dependencies.
Different from most GNN-based MTS that constructs graphs to model the spatial correlation between
variables (i.e. nodes) Bai et al. (2020); Wu et al. (2019; 2020), we attempt to build a supra-graph that
sheds light on the "high-resolution" correlations between any two variables at any two timestamps
and largely enhances the expressiveness on non-static spatial-temporal dependencies. Obviously,
the supra-graph will heavily increase the computational complexity of GNN-based model, then
a high-efficiency learning method is required to reduce the cost for model training. Inspired by
Fourier Neural Operator (FNO) Li et al. (2021), we transform graph convolutions in the time domain
to much lower-complexity matrix multiplication in the frequency domain by defining an efficient
Fourier Graph Shift Operator (FGSO). In addition, it is necessary to consider the multiple iterations
(layers) of graph convolution to expand receptive neighbors and mix the diffusion information in the
graph. To capture time-varying diffusion over the supra-graph, we introduce the edge-varying graph
filter, expressed as a polynomial of multiple graph shift operators (GSOs), to weigh the graph edges
differently from different iterations. We then reformulate the edge-varying graph filter with FGSOs in
the frequency domain to reduce the computational cost. Accordingly, we construct a complex-valued
feed forward network, dubbed as Edge-Varying Fourier Graph Networks (EV-FGN), stacked with
multiple FGSOs to perform high-efficiency multi-layer graph convolutions in the Fourier space.
The main contributions of this paper are summarized as follows:
•
We adaptively learn a supra-graph, representing non-static correlations between any two variables
at any two timestamps, to capture high-resolution spatial-temporal dependencies.
•
To efficiently compute graph convolutions over the supra-graph, we define FGSO that has the
capacity of scale-free learning parameters in the Fourier space.
•
We design EV-FGN evolved from edge varying graph filters to capture the time-varying variable
dependencies where FGSOs are used to reduce the complexity of graph convolutions.
•
Extensive experimental results on seven MTS datasets demonstrate that EV-FGN achieves state-of-
the-art performance with high efficiency and fewer parameters. Multifaceted visualizations further
interpret the efficacy of EV-FGN in graph representation learning for MTS forecasting.
2 RELATED WORKS
2.1 MULTIVARIATE TIME SERIES FORECASTING
Classic time series forecasting methods are linear models, such as VAR Watson (1993), ARIMA
Asteriou & Hall (2011), state space model (SSM) Hyndman et al. (2008). Recently, deep learning
based methods Lai et al. (2018); Sen et al. (2019); Zhou et al. (2021) have dominated MTS forecasting
due to their capability of fitting any complex nonlinear correlations Lim & Zohren (2021).
MTS with GNN
. More recently, MTS have embraced GNN Wu et al. (2019); Bai et al. (2020); Wu
et al. (2020); Yu et al. (2018b); Chen et al. (2022); Li et al. (2018) due to their best capability of
modeling structural dependencies between variables. Most of these models, such as STGCN Yu et al.
(2018b), DCRNN Li et al. (2018) and TAMP-S2GCNets Chen et al. (2022), require a pre-defined
graph structure which is usually unknown in most cases. In recent years, some GNN-based works
Kipf et al. (2018); Deng & Hooi (2021) account for the dynamic dependencies due to network design
such as the time-varying attention Deng & Hooi (2021). In comparison, our proposed model captures
the dynamic dependencies leveraging the high-resolution correlation in the supra-graph without
introducing specific networks.
MTS with Fourier transform
. Recently, increasing MTS forecasting models have introduced the
Fourier theory into neural networks as high-efficiency convolution operators Guibas et al. (2022);
Chi et al. (2020). SFM Zhang et al. (2017) decomposes the hidden state of LSTM into multiple
2