1
Hypothesis Test Procedures for Detecting Leakage
Signals in Water Pipeline Channels
Liusha Yang, Matthew R. McKay, Xun Wang
Abstract—We design statistical hypothesis tests for performing
leak detection in water pipeline channels. By applying an appro-
priate model for signal propagation, we show that the detection
problem becomes one of distinguishing signal from noise, with
the noise being described by a multivariate Gaussian distribu-
tion with unknown covariance matrix. We first design a test
procedure based on the generalized likelihood ratio test, which
we show through simulations to offer appreciable leak detection
performance gain over conventional approaches designed in an
analogous context (for radar detection). Our proposed method
requires estimation of the noise covariance matrix, which can
become inaccurate under high-dimensional settings, and when the
measurement data is scarce. To deal with this, we present a second
leak detection method, which employs a regularized covariance
matrix estimate. The regularization parameter is optimized for
the leak detection application by applying results from large
dimensional random matrix theory. This second proposed ap-
proach is shown to yield improved performance in leak detection
compared with the first approach, at the expense of requiring
higher computational complexity.
Index Terms—Leak detection, hypothesis test, random matrix
theory.
I. INTRODUCTION
Leakage in water supply systems causes wastage of water
and energy resources, and poses public health risk due to
water pollution. Leaks may occur, for example, due to ag-
ing pipelines, corrosion, and excessive steady and/or unsteady
pressures in the system [1]. Thus, an effective leakage detec-
tion method is essential.
Most related research in this area has focused on the prob-
lem of leak estimation, for which the objective is usually to
estimate the location of the leak, assuming that a leak actually
exists in the pipeline. For this purpose, various transient-based
leak location estimation methods have been developed (e.g.,
[1–4]). In this work we address the related (but different)
problem of leak detection, by developing suitable statistical
hypothesis testing procedures. Despite being a natural detec-
tion approach, to our knowledge, hypothesis tests have yet to
be developed for leak detection in water pipeline systems.
Generally speaking, we develop data-driven approaches to
decide between the presence or absence of a leak in the pipeline,
and for the former case, return estimates of the leak param-
eters. The measured data corresponds to primary and sec-
ondary measurements of head differences at different frequen-
cies, taken from multiple sensors deployed at different loca-
tions along the water pipeline. Our tests are developed based
on a linearized transient wave model in the frequency do-
main, as proposed in [3, 4], which has been supported by
experimental data [5]. By applying hypothesis testing theory
to this model, we find that, from a technical point of view, the
problem boils down to a binary classification problem that dis-
criminates between a “null hypothesis”, corresponding to zero-
mean complex Gaussian noise with non-trivial correlation, and
an “alternative hypothesis”, corresponding to a structured (de-
terministic) signal embedded within the Gaussian noise. For
the latter hypothesis, the deterministic signal is a function of
the leak parameters, including size and location.
Since the signal and noise model parameters (i.e., noise
covariance, leak location and size) are all unknown, we de-
velop test procedures based on the generalized likelihood ratio
test (GLRT) [6], which constructs a likelihood ratio based
on the two hypotheses, and replaces the unknown parameters
in the likelihood functions by appropriate estimates. We first
consider a traditional strategy of which replaces the unknown
parameters by their maximum likelihood estimates (MLE),
and develop a suitable test statistic. This statistic exploits the
known structure of the leak signals (under the alternative hy-
pothesis), and is proven to have the desirable property of being
a constant false alarm rate (CFAR) statistic; meaning that a de-
tection threshold can be specified which achieves a fixed false
alarm probability, regardless of the model parameters. Through
simulations, we demonstrate the good performance of the pro-
posed method in detecting leaks, and show enhancement over
methods that have been developed for related models in the
context of radar detection. This approach is particularly suited
to “data rich” scenarios, where the MLEs provide accurate
parameter estimates.
One limitation of the proposed approach is that for high
dimensional settings when the number of frequency domain
measurements and/or the number of sensors is large, the num-
ber of parameters to estimate is also large. This is particu-
larly the case for the noise covariance matrix, and it is well
known that under high dimensional settings that the MLE –
corresponding to the conventional sample covariance matrix
(SCM) estimate – is particularly inaccurate. This, in turn,
can degrade the performance of the proposed leak detection
algorithm. To deal with this potential problem, we propose
a second detection algorithm that seeks to design a robust
covariance estimation solution which is suitably optimized for
the task of leak detection, under high dimensional settings. The
approach is to replace the SCM with a regularized version
(termed RSCM) in the GLRT statistic, and to optimize the
regularization parameter to maximize the leak detection accu-
racy subject to a prescribed false alarm criteria. The RSCM
is a simple but effective covariance matrix estimator to deal
with problems of sample deficiency and high dimensionality
by pulling the spread sample eigenvalues toward their grand
arXiv:2210.13032v1 [eess.SP] 24 Oct 2022