
Table 1: Training on CIFAR-10 using a ResNet110 classifier on a single Nvidia V100 GPU. State-of-
the-art robustness training methods significantly slow down training compared to standard training.
METHOD TRAINING SLOWDOWN FACTOR
SMOOTHADV 46.20×
MACER 20.86×
SMOOTHMIX 4.97×
3 Maximizing Certified Radius via Knowledge Transfer
Although prior works have proposed methods for increasing the certified radius of the smooth
classifier, their training overhead is significant, making them much slower than standard training.
As we show in Table 1, training a certifiably robust ResNet110 classifier to convergence using
SmoothAdv, MACER, and SmoothMix is
46.20×
,
20.86×
, and
4.97×
slower, respectively, compared
to training a non-robust classifier with standard training.
Given constant innovations in architecture design (Figure 1) and the influx of new data, which may
result in various tweaks to deployed networks that elicit retraining, the large overhead of state-of-the-
art robustness training methods makes preserving certified robustness across model re-deployment
difficult. Therefore, we propose Certified Robustness Transfer (CRT), a training method that improves
the usability of certified robustness training methods by dramatically reducing their training overhead
while preserving the certified robustness. Given the base classifier of a pre-trained certifiably robust
smooth classifier, we leverage the knowledge transfer framework to guide the training of a new base
classifier (and its associated robust smooth classifier).
1
In this section, we describe our method and
provide theoretical justification for its effectiveness.
3.1 Transferring Certified Robustness
From Equation 4, it follows that training the base classifier to maximize
Pη(fθ(x+η) = y)
for
any given input
x
will result in the maximization of the certified radius associated with the smooth
classifier, provided Equation 3 is satisfied. Thus, for the base classifier
fθ(x)
, our goal is to maximize
the following quantity over the training set:
n
X
i=1
Eη1[fθ(xi+η) = yi]≈
n
X
i=1
Eη[zyi
θ(xi+η)] (5)
In the above equation, like prior works [
3
,
29
,
39
], we leverage the fact that the softmax output of
a classifier can be treated as a continuous and differentiable approximation of its
arg max
output.
Methods like SmoothAdv [
29
], MACER [
39
] and SmoothMix [
15
] that target
`2
certifiable robustness
propose training objectives that maximize this term.
Now, suppose we have a pre-trained base classifier
fφ
. It follows that
Eη[zy
φ(x+η)] ≥0
. Through
straightforward algebraic manipulations (see Appendix A), we derive the following lower bound:
n
X
i=1
Eη[zyi
θ(xi+η)] ≥ −
n
X
i=1
Eη[zyi
φ(xi+η)−zyi
θ(xi+η)] (6)
That is to say that, for a given input
xi
, if we minimize the difference between the softmax outputs
of the teacher and the student (
fφ
and
fθ
) corresponding to the correct label
yi
, we will maximize
Equation 5 for the student. However, to ensure that the student has a non-trivial certified radius, we
must also ensure that Equation 3 is satisfied. If we assume that Equation 3 holds true for the teacher
(i.e., the base classifier of a certifiably robust smooth classifier), this condition can also be achieved
for the student by matching the overall softmax output of the student to that of the teacher.
3.2 Certified Robustness Transfer (CRT)
Based on the previous discussion, we now describe our method for training a certifiably robust
classifier through knowledge transfer. First, we obtain a pre-trained base classifier
fφ
, which has
been trained using a randomized smoothing based robustness training method as this maximizes
1
If no pre-trained classifier is available, we first train an architecture of lower complexity (i.e., fast to train)
compared to the target architecture (Section 5.1).
4