(in adversarial examples) or the spurious correlation between labels and background information (in
OOD generalization). In other words, AT enables DNNs to make predictions using intrinsic features
rather than spurious features.
A potential problem, however, is that existing AT methods ignore the specific design of perturbations
when used for solving OOD generalization problems. They usually simply conduct sample-wise AT
[
6
], which only brings limited performance improvement to OOD generalization. The essential reason
for the failure of this type of approach is that the perturbations it uses cannot distinguish invariant and
spurious features. As a result, it improves the robustness at the expense of the decreasing standard
accuracy [
7
]. Moreover, we empirically find that when adapting Universal AT (UAT [
8
]) to OOD
problems, i.e., conducting AT with domain-wise perturbations, it shows stronger input-robustness
when facing larger-scale perturbations compared to the sample-wise AT (see Section 3.2). Since
the sample injected with large-scale perturbations can be regarded as OOD samples [
5
], we draw
inspiration from this phenomenon that AT with universal (low-dimensional) structures can be the
key to solving OOD generalization. Therefore, we propose to use structured low-rank perturbations
related to domain information in AT, which can help the model to filter out background and style
information, thus benefiting OOD generalization. We make the following contributions in our work:
•
We identify the limitations of sample-wise AT on OOD generalization through a series of
experiments. To alleviate this problem, we further propose two simple but effective AT
variants with structured priors to improve OOD performances.
•
We theoretically prove that our proposed structured AT approach can accelerate the conver-
gence of reliance on spurious features to 0 when using finite-time-stopped gradient descent,
thus enhancing the robustness of the model against spurious correlations.
•
By conducting experiments on the DomainBed benchmark [
9
], we demonstrate that our
methods outperform ERM and sample-wise AT on various OOD datasets.
2 Related Work
Solving OOD Generalization with AT.
According to [
3
], the performance of deep models is sus-
ceptible to small-scale perturbations injected in the input images, even if these perturbations are
imperceptible to humans. Adversarial training (AT) is an effective approach to improve the robust-
ness to input perturbations [
4
,
10
,
11
]. However, many recent works have begun to focus on the
connection between AT and OOD due to the fact that OOD data can be regarded as one kind of
large-scale perturbation. These works seek to exploit the robustness provided by AT to improve OOD
generalization. For instance, [
6
] applied sample-wise AT to OOD generalization. They theoretically
found that if a model is robust to input perturbation on training samples, it also generalizes well on
OOD data. [
5
] theoretically established a link between the objective of AT and the OOD robustness.
They revealed that the AT procedure can be regarded as a heuristic solution to the worst-case problem
around the training domain distribution. Nevertheless, the discussion of [
6
] and [
5
] is restricted to
the framework of using Wasserstein distance to measure the distribution shift, which is less practical
for the real-world OOD setting where domain shifts are diverse. Additionally, they only studied
the case of sample-wise AT and did not further investigate the effect of different forms of AT (not
sample-wise) on OOD performance. Other works such as [
12
] focus on the structure design of
the perturbations. They used multi-scale perturbations within one sample, but they did not exploit
the universal information within one training domain. In our work, we focus on real-world OOD
scenarios where there are additional clues lying in the distribution shifts, i.e, the low-rank structures
in the spurious features (such as background and style information) across one domain. We further
design a low-rank structure in the perturbations to specifically eliminate such low-rank spurious
correlations.
OOD Evaluation Benchmark.
The DomainBed benchmark [
9
] provides a fair way of evaluating
different state-of-the-art OOD methods, which has been widely accepted by the community. By
conducting rigorous experiments in a consistent setting, they revealed that many algorithms that
claim to outperform previous methods cannot even outperform ERM. Unlike previous works using
AT to address OOD generalization, such as [
6
] and [
5
], we adopt the Domainbed benchmark for a
fair comparison of our approach with existing state-of-the-art methods in this paper.
2