
2
ity of the neural networks to process data and detect
the GWs orders of magnitude faster, it can be effectively
combined with the matched-filtering based detections to
enhance the confidence of a given detection and also to
efficiently target data segments with positive detections
for a subsequent matched-filter based analysis.
Several ML based analysis of GW data have appeared
in the literature over the past few years, beginning with
glitch classification and subtraction [19–24]. DL methods
are first applied to direct detection of GWs by George
and Huerta using simulated aLIGO noise [18] and fur-
ther extended to real aLIGO noise, resulted in detect-
ing the presence of the first GW, GW150914 [25] in the
data stream. This work has inspired several attempts
to identify and locate GWs in real aLIGO data [26–31]
and many other ML based studies focused on parame-
ter estimation of real GWs [32–35] followed. Denoising
GWs using DL was applied in [36], and the proposed
denoising scheme was able to extract four GW events
(three from Hanford, one from Livingston) with high sig-
nal overlaps. In [37,38] the authors proposed a denois-
ing auto-encoder, based on Recurrent Neural Networks
(RNN) to denoise GWs. Most recently, [39] used 1D
CNN in an auto-encoder configuration to denoise many
of the GWs (three events from O1 and three events from
O2), using whitened data from both the detectors.
Although George and Huerta [18] in the foundational
article suggested that using 2D data in the context of
GW detection is sub-optimal because of the extremely
weak signal strength characteristic of GWs, we are using
2D representation of the data in order to separate signal
from the noise. In our denoising scheme, we use 2D CNN
architecture on raw detector data, with un-whitened sig-
nal templates and noise. This approach, to the best of
our knowledge is the first attempt to detect and denoise
GWs from raw data in 2D representation.
In this paper we present a DL framework to detect and
denoise GWs from raw strain data from the aLIGO de-
tectors. The detector noise at both aLIGO detectors are
highly non stationary and non-Gaussian, with very high
noise dominating both ends of the frequency spectrum
[40]. While ambient seismic noise (from ocean, traffic,
earthquakes, etc.) dominate the low frequency end of the
noise [41], photon shot noise dominate the high frequency
regime. The real GW signals are extremely weak and are
deeply buried inside the detector noises and occupy the
same frequency band as the detector noise, making it
impossible to separate from the noise using traditional
filtering methods.
In this method, we implement a time-frequency denois-
ing technique used in seismic denoising, which success-
fully separated earthquake signals from ambient noise in
2019 [42] using DL. The noisy strain data is first trans-
formed into time-frequency domain using Short Time
Fourier Transform (STFT), rendering the one dimen-
sional time series data into two dimensions, represented
by Fourier coefficients. In the Fourier domain, the coef-
ficients associated with noise are attenuated to enhance
the signal coefficients. These modified Fourier coefficients
are then transformed back in to the time domain using
inverse Fourier transform and thereby reconstructing the
GW chirp signal in the time domain. The underlying
idea is to promote a sparse representation of the sig-
nal in the time-frequency domain where the signal can
be represented by a sparse set of features which makes
the separation of the signal from the noise easier in the
Fourier domain. This method is especially useful when
the GW signal and the detector noise occupy the same
frequency band, making the filtering techniques virtually
impractical.
II. METHOD
The key here is to find a mapping function that can
appropriately find a threshold to suppress the coeffi-
cients corresponding to the noise in the time-frequency
domain, hence enhance signal separation. These func-
tions are highly non-linear and are hence difficult to con-
struct mathematically for the the GW problem. That is
where DL techniques can be incredibly effective, which
learns to build a high dimensional, non-linear mapping
function from the data alone during the training process.
This Neural-Net learns the sparse representation of the
data in the time-frequency domain, and builds a high-
dimensional, non-linear mapping function which maps
these representations into two masks, one for the GW
signal and another one for all the noises.
The raw GW data from the detector d(t) undergoes
STFT and is represented in the time-frequency domain as
D(t, f), which is a combination of the GW signal S(t, f)
and all the noises N(t, f),
D(t, f) = S(t, f) + N(t, f).(1)
The idea is to construct the mapping functions that can
successfully map the detector data into a representation
of the signal and a representation of the noise separately.
This mapping can be accomplished through a soft thresh-
olding in the sparse representation where the threshold is
estimated by assuming a Gaussian distribution of noise
[43].
From here we construct two individual masks MS(t, f )
and MN(t, s) which act as the mapping functions for the
signal and noise respectively, and are given by,
MS(t, f) = 1
1 + |N(t,f)|
|S(t,f)|
,(2)
MN(t, f) =
|N(t,f)|
|S(t,f)|
1 + |N(t,f)|
|S(t,f)|
.(3)
These masks are the targets for the supervised learning
problem, and they are constructed during the training