
Adversarial Pretraining of Self-Supervised Deep Networks: Past,
Present and Future
GUO-JUN QI, Laboratory for Machine Perception and Learning, USA
MUBARAK SHAH, University of Central Florida, USA
In this paper, we review adversarial pretraining of self-supervised deep networks including both convolutional neural
networks and vision transformers. Unlike the adversarial training with access to labeled examples, adversarial pretraining
is complicated as it only has access to unlabeled examples. To incorporate adversaries into pretraining models on either
input or feature level, we nd that existing approaches are largely categorized into two groups: memory-free instance-wise
attacks imposing worst-case perturbations on individual examples, and memory-based adversaries shared across examples
over iterations. In particular, we review several representative adversarial pretraining models based on Contrastive Learning
(CL) and Masked Image Modeling (MIM), respectively, two popular self-supervised pretraining methods in literature. We
also review miscellaneous issues about computing overheads, input-/feature-level adversaries, as well as other adversarial
pretraining approaches beyond the above two groups. Finally, we discuss emerging trends and future directions about the
relations between adversarial and cooperative pretraining, unifying adversarial CL and MIM pretraining, and the trade-o
between accuracy and robustness in adversarial pretraining.
CCS Concepts: •Computing methodologies →Computer vision representations;Unsupervised learning.
Additional Key Words and Phrases: adversarial pretraining, contrastive learning, masked image modeling, memory-free vs.
memory-based adversaries, instance-wise perturbations
ACM Reference Format:
Guo-Jun Qi and Mubarak Shah. 2022. Adversarial Pretraining of Self-Supervised Deep Networks: Past, Present and Future. 1,
1 (October 2022), 21 pages. https://doi.org/10.1145/nnnnnnn.nnnnnnn
1 INTRODUCTION
Adversarial pretraining aspires to learn an unsupervised deep networks without access to labels. In contrast,
adversarial training in literature [
11
,
31
,
48
,
53
,
55
,
58
,
65
,
70
,
77
,
82
] seeks to nd worst-case adversarial examples
and use them to train neural networks robust to the corresponding attacks. While some ndings [
65
,
70
] revealed
that the adversarially trained networks can be robust to adversarial attacks or gain improved standard accuracy,
however, there are limited reviews of adversarially pretrained networks by classifying and evaluating existing
approaches, assessing their advantages and shortcomings, as well as charting future directions.
At the start, we want to clarify a common misunderstanding about the role of adversarial pretraining. The goal
of an adversarial approach is not limited to learning robust representation against potential attacks. Instead, it is
also employed to improve the generalization accuracy in downstream tasks, especially when the adversarial model
attacks on feature levels [
40
,
43
,
61
,
62
] rather than on raw inputs (e.g., image pixels) of individual instances (aka
instance-wise attacks [
39
,
42
,
45
]). When not attacking on the raw inputs, the adversarial pretraining often cares
Authors’ addresses: Guo-Jun Qi, guojunq@gmail.com, Laboratory for Machine Perception and Learning, 10940 NE 33RD PLACE, SUITE
202, Bellevue, Washington, USA, 98004; Mubarak Shah, University of Central Florida, 4328 Scorpius St., Orlando, Florida, USA, 32816,
shah@crcv.ucf.edu.
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that
copies are not made or distributed for prot or commercial advantage and that copies bear this notice and the full citation on the rst
page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy
otherwise, or republish, to post on servers or to redistribute to lists, requires prior specic permission and/or a fee. Request permissions from
permissions@acm.org.
©2022 Association for Computing Machinery.
XXXX-XXXX/2022/10-ART $15.00
https://doi.org/10.1145/nnnnnnn.nnnnnnn
, Vol. 1, No. 1, Article . Publication date: October 2022.
arXiv:2210.13463v1 [cs.LG] 23 Oct 2022