
1.2 Related Work
Learning with Membership and Equivalence Queries. Membership and equivalence queries (MQ
and EQ, respectively) have been widely used in learning theory. Membership queries allow the learner to
query the label of any point in the input space
X
, namely, if the target concept is
c
,
MQ
returns
c
(
x
)when
queried with
x∈ X
. Recall that, in the probably approximately correct (
PAC
) learning model of Valiant
(1984), the learner has access to the example oracle
EX
(
c, D
), which upon being queried returns a point
x∼D
sampled from the underlying distribution and its label
c
(
x
), and the goal is to output
h
such that with
high probability
h
has low error.
2
The
EQ
oracle takes as input a hypothesis
h
and returns whether
h
=
c
,
and provides a counterexample
z
such that
h
(
z
)
̸
=
c
(
z
)when
c̸
=
h
. The seminal work of Angluin (1987)
showed that deterministic finite automata (DFA) are exactly learnable with a polynomial number of queries
to
MQ
and
EQ
in the size of the DFA. Many classes were then shown to be learnable in this setting as well as
others, see e.g., (Bshouty, 1993; Angluin, 1988; Jackson, 1997). Moreover, the
MQ
+
EQ
model has recently
been used to extract weighted automata from recurrent neural networks (Weiss et al., 2018, 2019; Okudono
et al., 2020), for binarized neural network verification (Shih et al., 2019), and interpretability (Camacho and
McIlraith, 2019). But even these powerful learning models have limitations: learning DFAs only with
EQ
is hard (Angluin, 1990) and, under cryptographic assumptions, they are also hard to learn solely with the
MQ
oracle (Angluin and Kharitonov, 1995). It is also worth noting that the
MQ
learning model has been
criticized by the applied machine learning community, as labels can be queried in the whole input space,
irrespective of the distribution that generates the data. In particular, (Baum and Lang, 1992) observed that
query points generated by a learning algorithm on the handwritten characters often appeared meaningless to
human labelers. Awasthi et al. (2013) thus offered an alternative learning model to Valiant’s original model,
the PAC and local membership query (
EX
+
LMQ
) model, where the learning algorithm is only allowed to
query the label of points that are close to examples from the training sample. Bary-Weisberg et al. (2020)
later showed that many concept classes, including DFAs, remain hard to learn in the EX +LMQ.
Existence of Adversarial Examples. It has been shown that, in many instances, the vulnerability of
learning models to adversarial examples is inevitable due to the nature of the learning problem. The majority
of the results have been shown for the constant-in-the-ball notion of robustness, see e.g., (Fawzi et al., 2016,
2018a,b; Gilmer et al., 2018; Shafahi et al., 2019; Tsipras et al., 2019). As for the exact-in-the-ball definition
of robustness, Diochnos et al. (2018) consider the robustness of monotone conjunctions under the uniform
distribution. Using the isoperimetric inequality for the boolean hypercube, they show that an adversary that
can perturb up to
O
(
√n
)bits can increase the misclassification error from 0.01 to 1
/
2. Mahloujifar et al.
(2019) then generalize this result to Normal Lévy families and a class of well-behaved classification problems
(i.e., ones where the error regions are measurable and average distances exist).
Sample Complexity of Robust Learning. Our work uses a similar approach to Cullina et al. (2018),
who define the notion of adversarial VC dimension to derive sample complexity upper bounds for robust ERM
algorithms, with respect to the constant-in-the-ball robust risk. Montasser et al. (2019) use the same notion
of robustness and show sample complexity upper bounds for robust ERM algorithms that are polynomial in
the VC and dual VC dimensions of concept classes, giving general upper bounds that are exponential in the
VC dimension–though they sometimes must be achieved by an improper learner. Ashtiani et al. (2020) build
on their work and delineate when proper robust learning is possible. On the other hand, (Khim et al., 2019;
Yin et al., 2019; Awasthi et al., 2020) study adversarial Rademacher complexity bounds for robust learning,
giving results for linear classifiers and neural networks when the robust risk can be minimized (in practice,
this is approximated with adversarial training). Viallard et al. (2021) derive PAC-Bayesian generalization
bounds for the averaged risk on the perturbations, rather than working in a worst-case scenario. As for the
exact-in-the-ball definition of robustness, Diochnos et al. (2020) show that, for a wide family of concept
classes, any learning algorithm that is robust against all
ρ
=
o
(
n
)attacks must have a sample complexity
2
This is known as the realizable setting. It is also possible to have an arbitrary joint distribution over the examples and
labels, in which case we are working in the agnostic setting.
3