
Bayesian Repulsive Mixture Modeling with Mat´
ern Point Processes A PREPRINT
approach is still sensitive to any misspecification of the form of the individual components. Another approach is to use
more flexible (e.g. nonparametric) densities for each component of a mixture model [Gassiat,2017], though once again
this raises problems with model specification, identifiability and computation. A more modern approach is to directly
address the problem of overlapping clusters, enforcing diversity through repulsive priors. Here, rather than being
sampled independently from the base measure, component locations are jointly sampled from a prior distribution
that penalizes realizations where components are situated too close to each other. Such priors typically draw from
the point process literature, examples including Gibbs point processes [Stoyan et al.,1987] and determinantal point
processes [Hough et al.,2006,Lavancier et al.,2015]. Mixture models built on such priors have been shown to provide
simpler, clearer and more interpretable results, often without too much loss of predictive performance [Petralia et al.,
2012,Xu et al.,2016,Bianchini et al.,2018]. Nevertheless, they present computational challenges, since the repulsive
models often involve normalization constants that are intractable to evaluate.
In this work, we propose a new class of repulsive priors based on the Mat´
ern type-III point process. Mat´
ern point
processes are a class of repulsive point processes first studied in Mat´
ern [1960,2013]. More recently, Rao et al. [2017]
developed a simple and efficient Markov chain Monte Carlo (MCMC) sampling algorithm for a generalized Mat´
ern
type-III process (see section 2). In this paper, we bring this process to the setting of mixture models, using them as a
repulsive prior over the number of components and their locations. Treating the Mat´
ern realization as a latent, rather
than a fully observed point process, raises computational challenges that the algorithm from Rao et al. [2017] does
not handle. We develop an efficient MCMC sampler for our model and demonstrate the practicality and flexibility of
our proposed repulsive mixture model on a variety of datasets. Our proposed algorithm is also useful in Mat´
ern point
process applications with missing observations, as well as for mixture models without repulsion, as an alternative to
often hard-to-tune reversible jump MCMC methods [Richardson and Green,1997] to sample the unknown number of
components.
We organize this paper as follows. Section 2reviews the generalized Mat´
ern type-III point process, while section 3and
section 4outline our proposed Mat´
ern Repulsive Mixture Model (MRMM) and our novel MCMC algorithm. Section 5
discussed related work on repulsive mixture models, and we apply our model to a number of datasets in section 6.
2 Mat´
ern repulsive point processes
The Poisson process [Kingman,1992] is a completely random point process, where events in disjoint sets are inde-
pendent of each other. To incorporate repulsion between events, Mat´
ern [1960,2013] introduced three spatial point
process models that build on the Poisson process. The three models, called the Mat´
ern hardcore point process of type I,
II and III, only allow point process realizations with pairs of events separated by at least some fixed distance η, where
ηis a parameter of the model. The three models are constructed by applying different dependent thinning schemes
on a primary homogeneous Poisson point process. Despite being theoretically more challenging than the other two
processes, the type-III process has the most natural thinning mechanism, and supports higher densities of points. Rao
et al. [2017] showed how this can easily be generalized to include probabilistic thinning and spatial inhomogeneity.
Furthermore, Rao et al. [2017] showed that posterior inference for a completely observed type-III process can be
carried out in a relatively straightforward manner. These advantages make the generalized Mat´
ern type-III process
superior to the other two as a prior for mixture models. For simplicity, we will refer to this process as the Mat´
ern
process in the rest of this paper.
Formally, the Mat´
ern process is a finite point process defined on a space Θ, parameterized by a thinning kernel
Kη: Θ ×Θ→[0,1] and a nonnegative intensity function λΘ: Θ →[0,∞). We will find it convenient to decompose
the function λΘ(θ)as λΘ(θ) = ¯
λ·pΘ(θ), for a finite normalizing constant ¯
λ > 0and some probability density
pΘ(θ)on Θ. Simulating this process proceeds in four steps: (1) Simulate the primary process FΘ=θ1, . . . , θ|FΘ|
from a Poisson process with intensity λΘ(·)on Θ. (2) Assign each event θjin FΘan independent random birth-time
uniformly on the interval T= [0,1]. (3) Sequentially visit events in the primary process according to their birth-times
(from the oldest to the youngest) and attempt to thin (delete) them. Specifically, at step j, the jth oldest event (θ, t)
is thinned by each surviving older primary event (θ0, t0), t0< t with probability Kη(θ, θ0). (4) Write GΘand ˜
GΘfor
the elements of FΘthat survive and are thinned from the previous step, respectively. The set GΘforms the Mat´
ern
process realization.
For a hardcore Mat´
ern process (figure 1), the thinning kernel satisfies Kη(θ, θj) = 1kθ−θjk<η, where ηis the thinning
radius, so that thinning is deterministic: newer events within distance ηof a previously survived event are thinned with
probability 1. Other approaches are probabilistic thinning [Rao et al.,2017], where Kη(θ, θj) = ηp1kθ−θjk<ηR(with
ηp∈[0,1]), or the smoother squared-exponential thinning, where Kη(θ, θj) = exp(−kθ−θjk2
2η).Huber and Wolpert
2