
(a) (b)
Figure 2: (a) Wavelet decomposition for ERA5 reanalysis and (b) cone of influence for model input.
Step 1: Bias-correction of coarse GCM output
The wavelet coefficients of the coarse, biased
GCM output are mapped into a different set of wavelet coefficients whose statistics match those of
coarse-scale observations; specifically, ERA5 reanalysis [
19
]. The bias-correction step is formulated
as a semi-supervised learning problem whereby reanalysis data is used as target for all GCM ensemble
members. For each wavelet center, we employ an LSTM-based neural network whose input consists
of time series for that center and its nearest neighbors. To capture energy transfers across scales—a
landmark of turbulent flows—the input to the network also includes wavelet centers that belong to
coarser wavelet levels. This defines a cone of influence for the center under consideration (Figure
2b). From this input, the model outputs a time series whose statistics are required to match those
of reanalysis data. Once the models have been trained, the statistically-corrected time series are
assembled and transformed back from wavelet to physical space to produce bias-corrected realizations
of coarse-scale atmospheric dynamics.
Step 2: Reconstruction of small-scale features
The ensemble obtained from Step 1 has the same
resolution as the GCM on which it is based (typically, cells of size 250–500 km). To achieve a higher
resolution and be able to describe the local effects of extreme events, small-scale structures must be
added. This is done with a second machine-learning step whose task is to reconstruct the fine scales
(i.e., those below the GCM resolution) as a function of the coarse scales (i.e., those resolved by the
GCM). The algorithm is trained on ERA5 reanalysis data in a supervised setting and tested on the
debiased output from Step 1. Each wavelet center is assigned a TCN-based model that operates on a
multi-scale cone of influence. Wavelet levels can be trained in parallel, but this might lead to error
accumulation across levels at testing time. To combat this, we employ a sequential strategy akin to
imitation learning [
20
] whereby the input to each model is replaced with model predictions from
previous levels. This improves robustness of the predictions, especially for the coarser levels.
Statistical loss functions
In Step 1, the models are trained using statistical loss functions; this is
a key innovation of our framework. We consider two classes of statistical losses: quantile losses,
which capture heavy tails and extreme values; and spectral losses, which capture cross-correlations
between wavelet centers. The former guarantee that the frequency and amplitude of extreme weather
events in the output is consistent with reanalysis data. The latter guarantee space–time coherency of
the output with respect to wave propagation and transport phenomena. These losses are computed
over several decades of weather data to ensure convergence of the statistics with respect to sample
size (see Section A.2). Crucially, enforcing statistical consistency with reanalysis data still allows for
variability of events in the output, especially the most extreme ones. In Step 2, statistical losses are
used as regularizers, with the MSE loss (or a variant of it that emphasizes extreme values [
21
]) being
the main driver for the optimization.
3