
2
II. METHOD
A. Existing strategies for enhanced bump hunts
According to the Neyman-Pearson lemma [15], the
provably optimal anomaly score for any model-agnostic
search would be:
R(x) = pdata(x)
pbg(x)(1)
where pdata(x) and pbg(x) are the probability densities
of the data and the background respectively. Of course,
in practice we never have access to this likelihood ratio,
since the probability densities of data and background are
in general intractable. At best, one could hope for a large
number of samples drawn from the data and true back-
ground distributions; then one could approximate R(x)
with a classifier trained on these samples. We will refer
to this approximation of (1) as the “idealized anomaly
detector” throughout.
Since it is generally not possible to draw samples from
the true pbg(x) in a realistic anomaly search scenario, we
can at best approximate this idealized case either with
simulations or in a data-driven way. The focus here will
be on the latter strategy.
The challenge then is to obtain a high-quality estimate
for pbg(x) from data, e.g. by interpolating from sidebands
(SB) in minto a signal region (SR), and use weak super-
vision to obtain an anomaly score R(x). As long as a
cut on R(x)> Rcdoes not sculpt the mdistribution,
one can combine this cut with the 1D bump hunt in m
to greatly enhance the significance of the signal over the
background.
In the original enhanced bump hunt method, called
CWoLa-Hunting [8], R(x) comes from a SR vs SB clas-
sifier. This works as long as the features xand mare
statistically independent in the background (i.e. the x
features are distributed identically in the SR and the SB
for the background). This also ensures that R(x)> Rc
will not sculpt the mdistribution. Using these proper-
ties, the full enhanced bump hunt search strategy using
CWoLa-Hunting was successfully demonstrated on toy
simulation data [8,16], and then implemented on actual
data by the ATLAS Collaboration in [17].
However, it can be challenging to ensure that xand m
are independent in the background. Even a small corre-
lation can degrade or destroy the sensitivity of CWoLa
Hunting to anomalies. This has motivated the develop-
ment of alternative approaches that are more robust to
correlations.
•In Anode [9], one learns pdata(x) and pbg (x) using
conditional density estimators trained on the data
with m∈SR and with m∈SB; the latter are
automatically interpolated in minto the SR, which
alleviates the problem with correlations between x
and m. It was shown in [12] that in the presence of
correlations between xand m, the signal sensitivity
of Anode is robust while that of CWoLa-Hunting
collapses.
•In Cathode [12], one learns pbg (x) using the SB
density estimator just as in Anode. However, in-
stead of the second SR density estimator (which
will be more difficult to learn as it must also capture
the tiny deviations from the smooth pbg(x) from
a small localized signal), one samples from pbg(x)
in the SR, and trains a classifier (as in CWoLa-
Hunting) between the data and the synthetic back-
ground samples. Cathode thereby captures the
best of both Anode and CWoLa-Hunting, achiev-
ing a signal sensitivity that is nearly optimal and
yet robust to correlations between xand m.
•Finally, the Curtains [13] protocol operates similar
to Cathode, with the main difference that condi-
tional invertible neural networks (cINNs) are used
to map background examples from the SB into the
SR.
B. The problem of background sculpting
So far, apart from CWoLa-Hunting, the majority of
the effort has been invested in exploring data-driven ap-
proaches to learn R(x) as accurately as possible from
sidebands, while much less attention has been paid to
the issue of background sculpting. However, signal sen-
sitivity is not the only component of a successful new
physics search; background estimation is also essential.
In the presence of correlations between xand min the
background events, one must also show that R(x), even
if ideal, does not sculpt the background mdistribution
around the signal region, which would prevent back-
ground estimation via the 1D bump hunt. See Fig. 1
for an illustration of such correlated input features.
Note that, in any complete enhanced bump hunt strat-
egy, two data-driven background estimations must take
place:
1. An interpolation of the learned pbg (x) from SB to
SR in order to construct R(x).
2. After cutting on R(x)> Rc, we proceed with the
usual 1D bump hunt: an interpolation in the m
distribution from SB to SR (e.g. by fitting a suitable
functional form to the data excluding the SR).
This work is concerned with ensuring the robustness
of the second estimation. We will demonstrate—using
both a simple analytic toy model and examples drawn
from the LHC Olympics 2020 R&D dataset [14]—that
in the presence of correlations between xand m, cutting
on the learned R(x) can result in significant sculpting of
the mdistribution. This can be understood by the fact
that R(x) must be a more-or-less smooth function of x,
so any correlations of mwith xwill be inherited by R(x).
Furthermore, R(x) was learned using events in the SR, so