2
transients from isolated neutron stars [
27
] or as
early alert systems for binary neutron star collisions
[
28
]. CNNs are widely used for pattern recognition
[
13
,
29
] and classification tasks [
30
–
32
]. Their pow-
erful capability to identify shapes and structures
has lead to the definition of Generative Adversarial
Networks [
33
], allowing to generate new samples by
learning the underlying distribution of the original data.
In Sec. I, we describe how glitches have been selected
to constitute the training set and how we highlight them
in the cross-correlated TF maps. Details about the
architecture of our classifier and the training method
are given in Sec. II. We then show the results of the
training in Sec. III. Section IV is dedicated to large
scale tests comparable to the analyses conducted during
burst searches. Future prospects and conclusions are
given in Sec. V.
I. METHODOLOGY
Our search for minute-long bursts is based on the
excess-of-power method [
34
]. We make use of correlated
spectrograms, also referred to as time-frequency (TF)
maps, as described in [
13
]. In order to distinguish
glitches from possible burst signals, we will train a
neural network to identify them in the spectrograms.
As both can be present in a single TF map, we need to
consider the following cases : (1) a glitch is present in
the map, (2) a burst signal is present in the map, (3)
both or (4) none of them show up in the spectrogram.
Accordingly we will build 4 different data sets to
include all the possible scenarios in the training phase.
The fourth scenario consists in building a dataset
with background TF maps. The data from Hanford
(H1) and Livingston (L1) from the first half of the third
observing run (O3a) are first whitened [
35
] prior to be
correlated. Using time-slides [
36
], we then generate
10000 spectrograms with a time resolution of 6 seconds
and a frequency resolution of 2 Hz. As the TF maps
span 1000 seconds and 2048 Hz, their size is 166
×
1025.
Since we aim to apply our classifier on ALBUS’ output,
the size of the TF maps is chosen to be identical to [
13
].
A. Chirp generation
A methodology to recognize minute-long burst sig-
nals using machine learning techniques with very few
assumptions has been proposed in our previous work
[
13
]. This approach consists in using the Scipy library
[
37
] to generate chirp signals in the time domain with
random parameters, covering the whole time-frequency
parameter space. Figure 1shows some examples of gen-
erated chirps. As has been shown [
13
], this allows to
train a neural network with no prior assumption on the
targeted signals while confidently identifying minute-
long burst models. Chirps are injected into noise with
9 levels of visibility, defined as :
V=X
i,j
Sij −Nij (1)
where
Nij
is a noise-only spectrogram and
Sij
refers
to the same spectrogram in which a signal has been
injected. The sum is carried over all the pixels (
i, j
) in
the map. The definition of the visibility is particularly
useful to ensure chirps to be visible in the TF maps,
preventing the network to be fooled during the training
phase. The visibility can also be seen as a measure of
the anomalousness of the input TF maps. We choose
9 intensity levels in order to cover a quite large inten-
sity range, as seen in Figure 15 in the Appendix. We
use this intensity criterion to build our second dataset,
containing 10000 samples.
B. Glitch selection
During the second observing run (O2), glitches
happened roughly at a rate of 1 every min in the
detectors [
38
]. Although it amounts to a considerable
volume of contaminated data, they barely show up in
cross-correlated spectrograms. Indeed, both glitches
have to fall into overlapping time bins while showing
a sufficiently high signal-to-noise ratio (SNR) and
sharing some frequency bandwidth. Even if these
conditions greatly reduce the amount of glitches that
contaminate our search, several thousands of glitches
can be found out of a couple of millions of TF maps
generated during the background searches.
To constitute our dataset with glitches, we need a
way to inject several glitch classes into time frequency
maps. However, the only tool that is currently available
to produce realistic glitches can only generate blip
glitches [
39
]. Blips are one the 23 classes that have
been characterized by Gravity Spy [
22
,
40
]. They
have a frequency between 0 and 256 Hz [
41
,
42
] which
would limit the detection bandwidth of the classifier if
used exclusively in our dataset. Therefore, we have to
rely on the glitches detected so far to constitute the
training set. We thus select glitches that have been
recorded by Gravity Spy during O3a [
43
]. We load
the data around the GPS time of the chosen glitch in
each single detector (Hanford H1 and Livingston L1)