
Question 1.1. Given access to a mixture of interventional distributions, under what conditions can one identify all
the intervention targets?1
Our Contributions: First, we model the situation of identifying hidden off-target interventions as the
problem of identifying individual components of a mixture of interventions. We assume an underlying
CBN and model interventions via the do() operator described above. Second, by constructing examples,
we show that, in general for a given CBN and an input mixture distribution, components of the mixture
might not be unique. Using this, we motivate the need for a mild positivity assumption (Assumption 3.2) on
the distribution generated by the CBN and a mild and reasonable exclusion assumption (Assumption 3.1) on
the structure of the intervention components present in the mixture. Third, we prove that, given access to a
CBN satisfying positivity and any input mixture having intervention components satisfying exclusion, such
intervention components generating the mixture can be uniquely identified from the mixture distribution.
Fourth, given oracle access to marginals of the distributions generated by the CBN and the mixture, our
identifiability proof gives an efficient algorithm to recover target components from an exponentially large
space of possible components. Finally, in Section 5, we conduct a simulation study to analyze the perfor-
mance of an algorithm (Algorithm 2) directly inspired from our identifiability proof, but with access to only
finitely many samples. Even though the goal of our paper is to prove identifiability of these intervention
targets, our simulation study indicates that our algorithm is promising in the realistic situation of finitely
many samples.
Related Prior Work: Recently [SWU20] considered the problem of causal discovery using unknown
intervention targets, and, as a crucial intermediate step, prove identifiability of these targets. They also de-
sign two algorithms UT-IGSP and JCI-GSP (based on the Joint Causal Inference framework in [MMC20]) to
recover these targets from data. As discussed in our motivation, in many real situations, such as [AWL18],
the off-target effects are themselves noisy and end up creating mixtures of multiple unknown interven-
tions. Since [SWU20] assumes separate access to each unknown intervention, their algorithm cannot be
used in our situation. Another line of work related to ours is the study of mixtures of Bayesian Net-
works. Perfect interventions i.e. do() operators on the CBNs create new interventional CBNs (Defini-
tion 1.3.1 in [Pea09]) and therefore the input mixture in our setup is actually a mixture of Bayesian Net-
works. This is a more general problem and was tackled first in [TMCH98]. They developed an Expectation-
Maximization (EM) based heuristic to find individual Bayesian Network components. However, they do
not investigate identifiability of the components. In our setting, we care about identifiability since the com-
ponents correspond to the unknown interventions. Along with recovering the individual components of a
mixture, there is also growing interest in developing techniques to understand conditional independence
(CI) relationships among the variables in the mixture data. For example, some recent works try to build
other graphical representations, from which the CI relationships in the mixture can be easily understood
( [Spi94,RSG11,Str19b,Str19a,SPU20]). Even though these new representations can identify some aspects
of the components, none of these works prove or discuss the uniqueness and identifiability of the compo-
nents, which is the main interest of our work. Finally, we would like to mention that the general area of
causal discovery and inference using different kinds of unknown interventions has received a lot of atten-
tion lately ( [EM07,SWU20,JKSB20,MMC20,RHPM15]). Even though many of these do not align with goal
of our paper, the growing interest in this area highlights seriousness of the issue of unintended stochasticity
in targeted interventions and the desire to design algorithms robust to them.
2 Preliminaries
Notation We use capital letters (e.g. X) to represent random variables and the corresponding lower case
letter xto denote the assignment X=x. The set of values taken by random variable Xwill be denoted
1from here on wards we use the terms targets and components interchangeably
2